var/home/core/zuul-output/0000755000175000017500000000000015144053423014526 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015144066636015504 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000301466315144066546020276 0ustar corecorefmikubelet.log_o[;r)Br'o-n(!9t%Cs7}g/غIs,r.k9GfB >+YI_翪|mvşo#oVݏKf+ovpZj!Kޒ/h3_.gSeq5v(×_~^ǿq]n>߮}+ԏbś E^"Y^-Vۋz7wH׋0g"ŒGǯguz|ny;#)a "b BLc?^^4[ftlR%KF^j 8DΆgS^Kz۞_W#|`zIlp_@oEy5 fs&2x*g+W4m ɭiE߳Kfn!#Šgv cXk?`;'`&R7߿YKS'owHF6":=3Ȑ 3xҝd){Ts}cZ%BdARO#-o"D"ޮrFg4" 0ʡPBU[fi;dYu' IAgfPF:c0Ys66q tH6#.`$vlLH}ޭA㑝V0>|J\Pg\W#NqɌDSd1d9nT#Abn q1J# !8,$RNI? j!bE"o j/o\E`r"hA ós yi\[.!=A(%Ud,QwC}F][UVYE NQGn0Ƞɻ>.ww}(o./WY<͉#5O H 'wo6C9yg|O~ €'} S[q?,!yq%a:y<\tunL h%$Ǥ].v y[W_` \r/Ɛ%aޗ' B.-^ mQYd'xP2ewEڊL|^ͣrZg7n͐AG%ʷr<>; 2W>h?y|(G>ClsXT(VIx$(J:&~CQpkۗgVKx*lJ3o|s`<՛=JPBUGߩnX#;4ٻO2{Fݫr~AreFj?wQC9yO|$UvވkZoIfzC|]|[>ӸUKҳt17ä$ ֈm maUNvS_$qrMY QOΨN!㞊;4U^Z/ QB?q3En.اeI"X#gZ+Xk?povR]8~깮$b@n3xh!|t{: CºC{ 8Ѿm[ ~z/9آs;DPsif39HoN λC?; H^-¸oZ( +"@@%'0MtW#:7erԮoQ#% H!PK)~U,jxQV^pΣ@Klb5)%L%7׷v] gv6دϾDD}c6  %T%St{kJ_O{*Z8Y CEO+'HqZY PTUJ2dic3w ?YQgpa` Z_0΁?kMPc_Ԝ*΄Bs`kmJ?t 53@հ1hr}=5t;nt 9:I_|AאM'NO;uD,z҄R K&Nh c{A`?2ZҘ[a-0V&2D[d#L6l\Jk}8gf) afs'oIf'mf\>UxR ks J)'u4iLaNIc2qdNA&aLQVD R0*06V۽棬mpھ*V I{a 0Ҟҝ>Ϗ ,ȓw`Ȅ/2Zjǽ}W4D)3N*[kPF =trSE *b9ē7$ M_8.Ç"q ChCMAgSdL0#W+CUu"k"圀̲F9,,&h'ZJz4U\d +( 7EqڏuC+]CEF 8'9@OVvnNbm: X„RDXfיa }fqG*YƩ{P0K=( $hC=h2@M+ `@P4Re]1he}k|]eO,v^ȹ [=zX[tꆯI7c<ۃ'B쿫dIc*Qqk&60XdGY!D ' @{!b4ִ s Exb 5dKߤKߒ'&YILұ4q6y{&G`%$8Tt ȥ#5vGVO2Қ;m#NS8}d0Q?zLV3\LuOx:,|$;rVauNjk-ؘPꐤ`FD'JɻXC&{>.}y7Z,).Y톯h7n%PAUË?/,z_jx܍>М>ӗom$rۇnu~Y݇̇TIwӜ'}׃nxuoỴRZ&Yzbm ]) %1(Y^9{q"4e?x+ [Vz;E|d1&ږ/0-Vb=SSO|k1A[|gbͧɇد;:X:@;afU=Sru CK >Y%LwM*t{zƝ$;ȾjHim @tBODɆj>0st\t@HTu( v e`H*1aK`3CmF1K>*Mk{_'֜dN${OT-n,'}6ȴ .#Sqη9]5zoX#ZVOy4%-Lq6dACYm*H@:FUф(vcD%F"i ' VVdmcOTKpwq.M?m12N[=tuw}opYG]2u<ΰ+a1tHayɒ aY(P*aaʨ@ΰ<pX X{k[%Egl1$9  ֲQ$'dJVE%mT{z`R$77.N|b>harNJ(Bň0ae3V#b,PY0TEu1L/]MTB4$`H6NI\nbǛ*AyA\(u|@ [h-,j7gDTÎ4oWJ$j!frH_HI\:U}UE$J @ٚeZE0(8ŋ ϓ{BpY]Q4`Iz_*2coT'ƟlQ.Ff!bpRw@\6"yr+i37Z_j*YLfnYJ~Z~okJX ?A?gU3U;,ד1t7lJ#wՆ;I|p"+I4ˬZcն a.1wXhxDI:;.^m9W_c.4z+ϟMn?!ԫ5H&=JkܓhkB\LQ"<LxeLo4l_m24^3.{oɼʪ~75/nQ?s d|pxu\uw?=QR -Mݞίk@)ƸXM3“QLL FN+\r]IrfWoۢ,"~CRlnbu;Bx`C:gev\g7 6܋ hH+P5co.Q/cC@.sABC{تI=;̶n2¦l::s@ -Mn3䦴mHЭj !'T9Xsl o:d lzzMvYź ^ ٲAPm쪊m\9htwmjQ\c5&,|^C.SS P󂏛o n8Fkb^s/&a[s~W &ɿ^\r\ߺnqZV@z%=\#|-3ڝa$ΫM|-LsXY r# v&讳YE 6X̀v"@L'aEN^8 n`т ti6{b?-X;|iDɈP͐b7jk *bmc`  SgkmOl7^~xAE,Pmqs;l};Щ۸l?28Ćn.I0Yhu ;ZeY}Qg?lvחzäTC 4zv)|Vy7߯@qC cN ͯ~1-b }kAn=)m 3fo˶_ XJNC5B~6%d+, Z`ͲH-nမ^WbPFtOfD]c9\w+ea~~{;Vm >|WAޭi`HbIãE{%&4]Iw Wjoru ݜmKnZ<X; ۢ( nx K8.|DXb +*598;w)zp:̊~;͞)6vnM!N5Cu!8Wq/`FUwWAֻ,Qu W@ Fi:K [Av*_958]a:pmQ&'ᚡmi@ zF(n&P;)_]µ!doR0`pl`~9Fk[ٺ+4Hhao-jϸ??R<lb#P-^39T|L /~p│x@Bq"M/lja\b݋af LnU*P(8W[U6WX ZoѶ^SH:K:%Qvl\b FqQI.ȨHWo;Nw$͹O$oEE-eq=.*Dp,V;(bgJ!gF)892sw*+{[or@x,))[o新#.͞.;=fc<)((b۲Eumw峛M2,V[cm,S~ AF~.2v?JNt=O7^r.@DEuU1}g$>8ac#sĢB\PIPfwJQJ;Qxm &GBf\ZA$Ba-z|A-I @x70 晪MV)m8[6-Te@`E|=U D(C{oVa*H7MQK"<O%MTTtx袥:2JޚݶKd7UZihRk71VDqiގ\<:Ѓ3"gJJčE&>&EI|I˿j2ǯɘCGOa9C1L ={fm&'^tigk$DA' elW@Tiv{ !]oBLKJO*t*\n-iȚ4`{x_z;j3Xh ׄ?xt.o:`x^d~0u$ v48 0_ | E"Hd"H`A0&dY3 ً[fctWF_hdxMUY.b=eaI3Z=᢬-'~DWc;j FRrI5%N/K;Dk rCbm7чsSW_8g{RY.~XfEߪg:smBi1 YBX4),[c^54Sg(s$sN' 88`wC3TE+A\ф\9ys8;ӷL:@۬˨vvn/sc}2N1DDa(kx.L(f"-Da +iP^]OrwY~fwA#ٔ!:*땽Zp!{g4څZtu\1!ѨW(7qZcpL)ύ-G~^rFD+"?_h)yh=x>5ܙQ~O_e琇HBzI7*-Oi* VšPȰһ8hBőa^mX%SHR Fp)$J7A3&ojp/68uK͌iΙINmq&} O L-\ n4f/uc:7k]4p8wWLeUc.)#/udoz$} _3V6UݎvxyRC%ƚq5Щ/ۅw* CVo-1딆~ZYfJ"ou1ϵ5E bQ2mOΏ+w_eaxxOq:ym\q!<'J[FJ,4N:=6. +;$v6"I7%#CLTLyi{+ɠ^^fRa6ܮIN ޖ:DMz'rx#~w7U6=S0+ň+[Miw(W6 ]6ȧyԋ4ԙ./_A9B_-Z\PM `iĸ&^Ut (6{\٢K 5XGU/m >6JXa5FA@ q}4BooRe&#c5t'B6Ni/~?aX9QR5'%9hb,dsPn2Y??N M<0YaXJ)?ѧ| ;&kEYhjo?BOy)O˧?GϧmI C6HJ{jc kkA ~u?u7<?gd iAe1YB siҷ,vm}S|z(N%Wг5=08`S*՟݃*־%NǸ*kb05 V8[l?W]^@G:{N-i bɵFWǙ*+Ss*iނL8GwMm[eG`̵E$uLrk-$_{$# $B*hN/ٟCZ]DaUS@''mhSt6"+ҶT M6rN+LxE>^DݮEڬTk1+trǴ5RHİ{qJ\}X` >+%ni3+(0m8HЭ*zAep!*)jxG:Up~gfu#x~ .2ןGRLIۘT==!TlN3ӆv%#oV}N~ˊc,_,=COU C],Ϣa!L}sy}u\0U'&2ihbvz=.ӟk ez\ƚO; -%M>AzzGvݑT58ry\wW|~3Ԟ_f&OC"msht: rF<SYi&It1!ʐDN q$0Y&Hv]9Zq=N1/u&%].]y#z18m@n1YHR=53hHT( Q(e@-#!'^AK$wTg1!H$|HBTf̢̋Ko(D*6"^eTh'4xpFڜe'fVQ7~'c L^ԯwIڣA.}H;Ë*׬=`^ 9]r鐃 -Dfi2|QwZk‹u^6DQ1&H凎c!n[mi3)WfsF:M"uҷs.1!뾧1%s,hQs|hx̗3%*v9(I;:'>uQ+v)vR/egBhAAdh]4H:nV$tHI98/)=mͭ ڐn}}~ק?g_6WĩDRc˛r͹=svAV.ߴDJnxkG:8@ZO'rCWw+J>=$ts,cJZڗOx2c6 .1zҪR "^Q[ TF )㢥M-GicQ\BL(hO7zNa>>'(Kgc{>/MoD8q̒vv73'9pM&jV3=ɹvYƛ{3iψI4Kp5 d2oOgd||K>R1Qzi#f>夑3KմԔ萴%|xyr>ķx>{E>Z4Ӥ͋#+hI{hNZt 9`b˝`yB,Ȍ=6Z" 8L O)&On?7\7ix@ D_P"~GijbɠM&HtpR:4Si גt&ngb9%islԃ)Hc`ebw|Ī Zg_0FRYeO:F)O>UD;;MY,2ڨi"R"*R2s@AK/u5,b#u>cY^*xkJ7C~pۊ ~;ɰ@ՙ.rT?m0:;}d8ۈ ݨW>.[Vhi̒;̥_9$W!p.zu~9x۾vC;kN?WƟ+fx3SuKQqxST Ζ2%?T74a{N8;lr`$pZds=3jwlL Eڲ t|*n8[#yN SrA GYb8ZIaʼn8 #fg3i`F#5N 3q_M]j 8E!@1vցP7!|+R@;HspSI]ڻCZUcg5pDcIϹ,oN-_XI,3\j ]ٟ5~' SuipA!C厐$&k7dmhz/#"݃,YqCL$ڲ`"MUbeT>Xuv~4Le͢ }UVM)[A`b}mcE]LCEg=2ȴcmZ?E*-8nhױ1xR2ϫCya` A y!?h!9yL%VLU2gr26A!4vbSG ]ꧧWp/ &ee *w$-`J\ ptǣC^p#_`{ К8EW>*(D{ٛ,[fnY𱹞M=6&$<,"lX-Ǐ_whaE 98 (oѢ/Р΅ 7ցl6618ł_1/=fu).s¯?.S[{'g=Ҥ):d8h\y6]t1T7IUV:;.1& ,5΀j:<< +Y?58In'bXIǣO{&V\DŽ0,9f O_"[l:h¢8wݓ19\:f6:+ .3}=uvKc ٹeS<>ij(o'ciS<{1$E[nP b?8E'xv[K+E{,Qƙ1*dcs_Z'407|qBOgYU|U--sG8`u! qGYܷw;ȌCPc_|(RaIBKb+{P.T! =ĦiTob d<>SHr][KqWs7ѝBYǭ~RR"p9dFg|K- obY_vM 4>/]e/dy,8!xŋ5 R<^mYo 3c9(F?h'$0&*m.)HzzBvU0h} -_.7^nya+Cs 6K!x^' ^7 l 2Jj.S֔(*CjaS:vp/N6I*x8"EȿQa[qVM/)fpOj4r!:V_IG^nILVG#A7jF};qPU嗈M9VS;a+Ӧ8E8zmMs*7NM~@6 ' 8jp*:'SOANa0rӍ?DT%l)gvN}JT(Ȋqm|dc+lQai,|Dߟ|, d#EjZܴv]pEO7}&gbXԈedKX :+Z|p8"81,w:$TiVD7ֶ]cga@>\X=4OZSܿ* %xccDa.E h :R.qɱMu$ơI8>^V Y. ,BLq~z&0o- ,BLqfx9y:9244ANb n\"X>Y`bb*h%)(*_Gra^ sh6"BzƾH( ."e)B QlKlXt҈t9՚$ضz]'.!-r"1MCĦʸ"66pE{ =CNc\ESD[T4azry !5yY~ :3;Y[Iȧ q:i Ǟ/"8Wxç,vܰtX-LE7 |-D`JLw9|fb>4Nu ߏ3ap5k_JA+A.A~ C~`[KaQ-Ģn9ѧf q:cT >to^ X]j?-ȇlCf0hM`~ ó}0W@o  K[{d+`ze"l |d;L2k%x90ݙ^Oe ]nHfS+.4<#/5߁ݛǪ0q,7FeV/!; 瓠 Li% z}ɯww"O-]J`sdN$@"J`Y13K/9`VTElsX|D^c%֯T][$m;ԝ!,Z5f`XFzȁ=nrSA8; P=uY}r/27OUa%~0;үM3Tu ȩ*'3IC~LG,?.?C3tBYpm_g.~>3ʄ55[c&-Wgy_jVo,?s w*n\7[cpMY<~/"˘oV܉T6nn \_ߋV_}Z=k-nn sn.*upw pX\_ U-C_wS!|q?E-S_w$-#9?mh{R 4ѭm_9p -h2 dֲ 1"j {]]Nk"䁖%5'32hDz O\!f3KX0kIKq"H~%.b@:Oec6^:V8FDza5H`:&Q5 ^hI8nʁu EA~V O8Z-mYO!tO֠υ9G`6qmJc,Qh: ݢKNw2taC0Z' O > f-`:F_Ѫ2)sCj1THɩhS-^p b~?.>, `0!E%ҏ:H =VՑӄ| Ć.lL t1]}r^nʂI-|i*'yW='W6M$oeB,޳X$I6c>EK# 15ۑO2Jh)8Vgl0v/eNEU"Ik dRu˜6Uǖ xs%P ع omWl҈sApX!^ Ɩgv{Xn|$̇d`>1Ljn떚F+B9l"UP۾u2Ja>0c0Vvގj$]p^M+f~@9{bOe@7ȱ^%u~-B竟} |23 Z.`oqD>t@N _7c$h3`lg\)[h+pHBr^J |r\8czEnv@qZbRT1e8V Scc6:$[|a.fpU`ZR֩bKgTlѩynۢ, "1LӰW&jDkM~# (C>ϭQ3{ߤ%EN;?P%ٱm -{2k 8Vbv"wŏݙmn&O1^'}plM)0\n ή ?Cֲa9H] lX9^vCο -vd+OUgRy2Я\ B0!% #>bJPUck\Ul'F瘏Y4Ew`[x٘p,>9V"R1I>bJ` UL'5m1Ԥ:t6I >jz(:W֪Ƹ)!fꠗe[XLE4atGS1px#S]MF˦NJPYDX%ܠꡗhl}i9f?q>b-E'V"mNf""ŦK9kǍ-vU #`uVi<s)/=r=nlӗЩsdLyVIUI':4^6& t,O669Ȁ,EʿkڍfC58$5?DX 4q]ll9W@/zNaZf% >Ę_"+BLu>'Ɩ=xɮ[⠋X((6I#z)2S zp&m?e8 "(O+:Y EaSD]<^(]В|Ǚ8"oRs?]\McZ0ϕ!1hKS`h0O{!L-w]ln2&0Ǚ'0=.T4G7! H/ͺ|@lX)+{{^s1V63 ۗI"*al NJ`Q8B\pup6_3XqCXznL9:{o qcuו8`n{ave=}OR9~yL Z1=W8>É R$|L ]OfJl˪VVg:lDԒ͢Zu[kWۗw{{7st08`J0ꨴU1|z:9dX)z2!S:'q9| 76"Q;D*04Zٚ ?V¼r8/G:T6Fw/ɚ~h?lUc3MکEen壹n\殸,˛_uu.Jssu/*47U0)l?R_^Uon̝f-nnZTeuu nn/*0׷տ·sHH?Et _I`[>>0ւcz/Adh.$@bѨLtT=cKGX nݔ͆!`c|Lu_~ǴT?crO e9d ljB?K_z>p%'3JQK-͗R>KkΤOq,*I|0]Sj%|-Ԟ = Ʃ%>H&t;9`>$& nIdE Ͻq*nŘʰNҁV_]ks۸+*/D+UZ?fr]7N;J@+|m,?X[w=sh4t;J"<ꨤ罪`poKi hDH"[.x=NVŸw n)--־3^(ʼ(U?3 :<<ϲ%#!g~$3e1m&FOOI$͛!OBOFb3C&\Q*<$dV](yDe> _l'y!e4;@3dOG(xT2E!@EHFը%OQ5˛'GJQUȰ=t>w"pΥ꺨OٮjR겚O ߒZWx؋h^NFQ^Rx8SW-puS&]R]'h=?+ ? 㬻fK12+oUyT}8IKsP!EM9D?Fgүiz8.y|\9Wl y%ǣ秼 ߡͣ ~x Zx&z<'/HWTvٞt{cY9#>uG=%h9#h8AGx`H#x zV33ꦞtKRHreSxJF2LӬo7ޏ> \rq.æxLr(R1/LIBqq ʎ:S=T "P,5s88;9Q (!os9Y\}8V EߜI6%TGݛϛ\`u9g:.o_tW CtT|3FA$C@ Uȳ6}]ٲԜM7(f$M'Ӓ ~;׍SY`E(B3l\uMGck:D~]6*=\FF :˔ggԀmh[Ě6rJBmT=d%ʼ/] Enx OP=o`:O>80ݮ>ͳKpX>&wnhU86 B '* R P\ lW-1\&Gg*|͚>8Qkh>TT29BJ7w|ggۨ"B}'AzsPxpe=9,ԇmzT6/ȓEW1d"A-)=g 6P>Kq )擾ZMQ/'3L.^C0A?m u7)²lLqڵYOQ^bvnpLG!S c!o Q\VeU Oj$ڴ#&iIw<ܓ{nVz,9q(CgqIO{2H P*xojֳfDtVq Ya_G<~d\e{ ա%I~%dH {,nҼ upO sEMb"G nXwV4[ǁa?V8: :Cr?2SzC.qw7\\q<.!E.8b6קߋ[쫢SrmHx{H3gGKH|? BuUCx":aC@V3DfE5:xaTԮK|HL=j"sѮ lzt\)n ttC|~mEHu{e=+WaakQ2hSPƌ{jpJWAI?BNFyJ#uQ YlC5DUp4eO;;FV 66쑐ozwW2ΒW/x@:Ku 0DK: Sֺ*.T&)Q<^vlf hkPv`efw{8d]"`p+Յ!Z؞KVOK BFW CIG\qf8s@s2bu C=[]_R2TI aZ:FNvQ)52S15E K-YTRBEv9O +x\h_MZWaJLA) 3z{#3q[ÃAV6W'T bj9iA֦Qb||if!u} Z_37r2o x$̴7%0D0b? Y< ,O lN;d6ňCeՕT֤r^1j"jEuxҨo p]+{&LWpti%!L׵hKɰI$mze~Ux4u%j]+@a\q'FӶ.;.u[bD45E0@!z(ij,`WHP_z6>ɫX엳0GP۷VxZ>IRLw)jhn&ԨBUQL;V0yڟ=z[oZZџ9aa4V,]8eI) bv>%AE>C,XXV\xɣ^e41@iӅ. LOpkjY^xX V%֋$\Qbvtn8P&h:dYrq c2d!k \%ʝ{;™Mm2{MܕvOS~Df-unsHx:խ 1ޢFXkA=sP.]@*WPA+u ]j-3sJ%yԮr01g=#zHKc3;ʈWd QY] B}ewTU}6Ow}־q=mj>%uRueF߂0j62zHWF\+.W;q)oѪu^I>DO@sR] shB9ę8֥\yD6_>hVYsc۸{fqk O'o|p ܪ+cگ~# sڥ A,{{j@eCN=/u7Cunߣm:B]s  Lwԇܖn@2)pav%K]]/iH!C,ri]a&* zqXt[s1f@f[ҙw*.ۂ% T"ml3bZ?K#w KANQ~]{|G|fS>KU}` p)X9kUz߼Ҩ~ew'{IZLHWO7jQDU6I9ص.Ve-&1tJu"o?ߵpᅬ?쟨o_ˣȿ20ʼnz" 2"w~pD8iBnt a8{xBo5䴙ٝ%:rL@[Ā#ٖ.`!'45/F,n``  @ Y!5B e/| (G%|i0p`^}=SKt6̍p Aw4&!3`L;" 0+KV=9-@/@,~}!y"747bvC('ň%| 7vL0"^8D `0X/tGE7  9# ާB:=K"F`> 2 "/1;u@(Ornľ&':/0"YĽt#bnlPCfzإ7Gx/xpQHfrܰ/9G |cW;f#q'?'YL f@ʰbn13$dPa!{A ˣin}Nݾ~x3 7n;syb~hД_gӞ 0$\~~. JDunbw; < pmnzF ح] ,]& zX%\g}ALKwI =;sp2t;W-J1"xYW%M5pƞ Ly 1?˳n Bv> N9*A"\ח?A z;sYHˡ~8PHK:uVo^K¨)xhC'9) uMza]jz,5 d MXm'`fY3KAܟ$jc"14gZ%*s ,Wqym1b &u&{0BqKw*$㲘kxI>֡`/yHۉu0Ϧpǟj7OhOQUnu`(JXg /*x+Fj̶%6lEj7Ԭg}eu\n;as#XӁrqg%\$eufSצ4bcxV 9vfgJ8αz:ʐZTM7gIY=Ⱦbd00\= S0lLl,]~=E0[$Õr B|wAEhR, x)TGL` 8;zB#sd;D$g )=f9Gl<ɺX,U40;: 7(_)%]򯳯>YI dNPހ ia_`/*"ҪRAk 'i";[uB *A.βMb\RR*k j 6,cf>{8VdTrNP, Z VHTwYpmkR6[o('FG9ql'.m=bn 7Q3hf496C qUv^Ҹ08~(F0hlKG3>}6]?2EDco\˯_L)ZOQC >]2t>!S!rV^:*#'7gg6ޝ\&w@ZOn ?$ 3X XH\Ck;HYom^dR3t] XՁ0Ԟ+n]vM" l|gʥZnzϞp"s,hP D;ҭ' 24`tpWe <Ք J[J>MPHA낺[n/4AG n!.zO{[o/4AG o!([m!(^P4A#e[[ʷ?MPHAuAOT Qz <\Z]ƽ+0YH H=_g>Xr\DcSH.5g mqo30Z.@yru\g#UdQzns)ymz=ea=Լ5/R̷bhLn rg elF& 10{Sk/@5'l8eVKƒ xX#A.w>ޔVl; BE<Ɠ_XPq t@8aP2N2 c,T%aPH R5Eƒ`d-4-;%l^k:R=)rT ݾA@Q?\Ͱ1xOZ@ 2Vɤ> 齂tLuǬuTGyZ˧9lvhr;07'2woKgea=k2 pA@G E\bf;Cs<S*wf\"7LppW#&{#L%xq{ P eKw 格"p~ f!dseD5PF:X{TZ賅+I}UPCWZzb;]hcxtlP^E` ~/8ApIo\,=jWrx";(Ϥ&b|5(׊Dd3'0قlDz`~u$=]\rp{4@vP֎ n vJ) g}n7eC_WUç^:g.upuF.\m(캏Tn)Mi-c2_{J*|Ϻ"Jɶ^uTLI U&wί:K2͙ BgwWs4OKU߶N3jf410m awl^!B#(-nHО ƕ]eIi x*r38WLѫ "81âx2Ƣ@ R i& a(:TYW5uJ&A$!%B!W LɁz)p ։ /]`"P@͑?!%zG7{[ 'ad&mAW`'j[B`E CH2kcѨ(-~b}x+9?qhl_Cdn*Ǟmq[^\럞YCkǒrw.Qwד4(8sS0&g#sY@G(lsAXaA3E%JIJ==1:ƹ~$i݂F! NtJĪE1iĖMJ=⯣MܐN)zNfiJ"n^3!Fiv- n.*-ЬVHǎ`^)N‘_bq?)$Ux@ 2W)b>) J۶?7΁s+6}TGlDC·FM;+]("+])ԑ nn*۽XH)炗TYXkfhsMK_$$Xȫ &ly2vÂ=yg.:@`X.A2Xh" n j 0A.KQ2ZHgTTdi<1Kt߶߳&imE4BC7 +4+{iWS$MVqN[|jd0RJW&'R$n 'M& 6m]`[s-9҈~uE%|iPQ潲 UI^%gqvJ0(Bn%v 9T2F )W-S"8lkSs.94Ta} ,8Nn_+|QMs>aQ --,B=)iI0F%TPSFʋ\fyo 0wYE cF xXmH:L22YdN譎J }Eg$rVVt"='20]5H*ڈ]#IWpQ' #(CXmEՖ^T[>Yg%JdQMW|s`lK# N#,5=wܑ׀2f*pY5e im,9CGPT![T&Y,7AҩaTí2\ %GWJeI8/Zq$8C`I;uY)CJ2zkx8gdiUCA.5 оU}11b_,A3d.j]l4^iSO՘0v|!Fkȍ!;cMh$uyA tYp۠amU$ 6UX%R$1m^Eroޑ׫3"@jLL R6Cc:^| c1uQG<^Xׯ\Nn]_)G9aCTy8*=ɫmj=Fjj-̭+eh0'!)Fc4&z$n^V.O"h4xQS #'K+Q$&G;y5CuRv5$%GՐy%I'82;ϭ6 4e:"N^SsniёN)<۫G8 wtđwt2Tϛny-BMvdQb( Ek#(Np0E֜S`'HHHݠZI *Ht:ib|,8Nrdvv&jd*AmX*H-EbjwEsF{2 F@'J^ E (;۲[CZơE'':7{]A 5cauZrww͇[GL\]e D/ud_v78Z8{;䋀ouW}UC('wbE@Ki̒\&&GLR$$7Meq%~c&Oze^-g.@lUH:mr͂[&S =@V :]HwMݸp?U63$'/9wʚ%1|sy(Q|YxiTE±OvoO,8NYh>>iRYݬQg+5 Am\@*a&j" #ÛUe2~'JWx4~}'rRSSt>3c^Tkkޫg鹰D{js L2pY {ˮ?;d}.Y8Ωfq2>C䐶!7б El? Onl&hX BEn Mİ.7۷/V/~_= K_^B_NPW9&V/R)04^7We-rEҩV78y}UErÂΆsGIEQiڨ'KѸ[1`(Jx4$V]=(/brmHO47jwZ)B:Dע IoG/*en54&\ԆCKxܶU9v%?wMײkJxKg|؎۫jug/MD/7u nem,f/w}zЏqEm^Dl鎘b؂f%D_?_6?4^>@)c|^Y߯4n?YvK=m3R-)fEc42.jn:AZ崾59'K5}2C,#RIk.:@t4W_8_xQb'%X/汳8y'7;{M5_dںqʂ:tEG-j&At׍{dqswpR,L7wk@6 lێw@5Y4NQxK?59~V֕tvn;M^4{nOAH |*'Im{ -i4蠉ᴳ҅"DWLt⊿^?nYp4棯BM) s7\ iq-5NN=eHSS+M{axwaS^[ \T[pQMjE)ٷ- ] 00+g!NrS }䮕F ~͸߇|׏,u}a* U|+lYpRPـ H?LAlS3K=Ft0߫a1}Q.GFˬ 9A)N:[aB)bƠԁ6E0ql#~N5&SˡS$"(b)bkG84ևpQ-5oON'G_!L|95aaBs-Rw5Kn:xQgU}*1)#-6 ?BZ?9gGm88qP˵EF,i{.t%Ɖqb\lNqQİ0:MF(t5S;@gP;I)>l\U\ WMNB Uzފ[$Y)Nxp_qi+6 ? Ex{ۏ.E@Fc,{PRbqVRC/ؒ8z8 9$i\}OI»vM/n󑟆h:e:K뱭Mu]1z.%$O; k%o[9 }AaߍYnFR^B V$9<=Nf;ޝ~b-ޝRDJ4q<:/< a胟'K0AZݓz f^9~"+sD+b*#!8ˇ]/N>糋l:7~NhR Q*W3LnBoUh|mhb:6UN LrП{ 5h,0k?69. \.%8/Y3~otpTd.8dii>T'k, c?dzJ`XSrĖ'Ii[ xV@,$.PU4 M|cŌ+DaA * :ʋYOތLQ @U1<: Q) !h8 }X &es|(g&@`h0Lf2W UxC,go^z^4qwCc"c~e|aB蒷զ%e4~63{ۭ?{lH#8ZXDH_{5_rldɧ9&d (ǃd۵/ 0XD~@](}]_2"+Z9iE\ i&eWiNXzp,Qw}.GB,cctrx(RWChiCehM'%*?> lP+Yy8Z eq_Tmޚivv!vl;%aaF]M_UN`? 767໌,t\}juRiRo^;mu( . ¼{xu4~_#^19x,"jk餴G'eA\D7cBVߪkvLxmc(.3E>bIvp0|6!Hծ~o~/ -X'Dc]5f!TT(V.dr+J 09R.Z2R@#Xd BFÐ*ž]hA.*D[wKWh˶>JJt wСi#8/H"FXUBE}s i ,S4nF9'#H*c>B;>2#m>)o9iMSd̏ ^͏ ]2^E. u+k4_4ț'ϼ&(PRUG4)_KX2UP8(=*S]tgYZOft<`]`]ctKް,V3P ? _4Ϯ}i˟J^QS SnA/_r_^v1)f*`DiY!LYp޳,IԻƦgM4[{4coa"VG68" dG`T,}Ӄ>k9w~ιYCЬ%ԉTaBfe:z؛NRҦK ީjf;,%RaeWQE;Vjr=sxc{z>{*sh&&_=i[,шs0;u^7{qXo5c DM`GL'0uS,PƝ3R:s r ˬ-~R)ŒaSCˢφž>i8젎gv;w:IbeWܸO8\{#;~B=@ kFi>!+.qqWMq>ɽuk& ^J+7ɇ`vazOjݵ+#JQ.w)VŨ6Jj"{˙r n"!` F\p/˥wT`~M#k3$t:K_boQ;ǕYH F~_M={JcE%/{\"@iZBk%EezFȗN!}ah@>A gprC o.xh_nT5UXo/oK̨e.'cM)X"Q tcIN>XNPjue =vμ&fx/:gcCe7zy4 n0S|X >7I2xEhZ_! &M^G {l5ў]jEL!-& f4:}E`Vy`y=m J9}(σt>0iF3K2/ O!}#JX+KGfd 2La &G8n;N|OlPy0x9.xh T㯳FKϑ֍ Σ֌Mt9k@m 7 lg:EZRY/ QS6JB2bkhP+h5Bv36ef#$t#$IBRa,$ SBۥj}dejkhP+hx ͐LtH36Ś!@uQ@`!!cɤZ8i01*#)._(c[kFۆZiF`;ddH;.;S#R "U*3I9`"<3,֚ѶVѶgT**~~ͪ~+f%K UrO70Psqr^iqW^ >}‡@Jaz*r?r|%qqhɻӿzFe>;}u3+` jRU9s _Z  gV>|?^Vl\ M88wތsqEoƹAxw"́:νA%&GAξ. +C| vmV}YEϞN9J)1j0m_J)ae`gQ'/K& Lûkr5O]'nuD%ep-g4BOg9(hz&у2 Y U^6$yp[ >th?oMq#{&* )""~7?SD? ޜLsPue''ͨ=2b>(w!/'Ͱ VssuÍeoa |s3FkoN7mjm$wMn 4 |mA['ݒ5^t6oh(tqdcMS@vҮT`;,i4N7As%DKxyOq\~32E(J s䫧slxZ*i;@ PrLSKCSq2ιSHG3Ju2`r RP*sia)RfDX4eEP8gBTGB+1ʤwD3TQqFQoQ֙T1I_REE˄ʵ-$Y>WDxpLxG%$X($sil CTyTjqFIne 3~{Ffo>s,;;p/3mkmLIn^qWLx|d 6v-s&KȀkC-aYmIsKÍ!x!U)wm>O._kP0i_Խ ?YT.چ6V=vL*6Apbgnkq# J"/SAuS jdS;GU` Xd5(o"x`ZViw0%?˕K#[Z*-#[AMjpEk4fRSWy?Pګoв/]osp%oki.D)zHQ$+\L 0_WKjpijZEng"֥DZ-pbuw>%jhV ,klЏK{Sl'O}Q P+-ͦn_=:D8)'1ߦ8éӫEq_/EYjYb& &S?K(K_֘J'-WHu/uzyзykpT,ۓxzgu'J ]ptqbL<&zRXlΕ~Jپg-ℴ.o~#[V̭3Xu*0WީX5 \U[NJ8WX%|!:D˪TWV-UeD@VQUVX'Ey(ۏ[aF8݂J5idux\TKŧD{HiQh4"XDU5<+&iܦJiC"<߿Q!ㅖD<,-VB{Lc4t5R`bDjiFXy}-\%(ۘoSjL|6ky+ۦAօ=5%wFLm_j-frU=Jbˍdv#6Å4! Pq[e& xWJ .1m o7ݮw6܋[lZ*vwj( Ӱg7>%>1>޳mMg=zRtƩ4 E_<+Q/q[~̾UN )0,M:_ۏԾw:KWo*C&I5Z1Ne篶{D;xBыCN3ec Ξ*vf# ty떞ٓ>ym5^jBP{tgteϠ:}h7`Yd#d)au{c55TFLH^U+ekċ0^Wg/([kg0Ke6H 0R=,iX_Re_D2@z2zsF,s]^zGLzpS^{UmJQ uդo7%WB"ݘ:p7kL[فʻ`CWdU3\r~C^Xv cO"iiњZ¯ /;j4Z΄g,.>d1oeei balyr!l$޳ZpK]NE-A?/<#='VZpZJ(vTh.Yw9lƹbp˳axZ4soځq{r>?a6  z_NQ! IV*I#ͻMlDW ,vTL<΋(M/YO^^\OI:dW&.]j>$C*=h\VVKa/% no+6X'&^ ؀7ޝ ) ?>KA9pTFXa~4Ue ;N.`:Ĥ-Q ʲ$E5/)WjM(*"1A~QIοO9y焽Ct[|6}g:!J3^E~g=0a{&T(:C=J!I,tq6erLzجK#@ZcnB,w{z+3OaR }}yjYҶ}:hAa^H/LmD"3X/aY^*!+_"|`T8eB)2`DR}X36\2H5|)yM{caplYaV":Qc{rȳ|o04w=]e+(;0?$x%q'kMC킂 MN`]"A׍9xcͤo8fid;c!HrcIgg#[PR=VWh0{0=#XbG^_1WU|U`(<jڨ7v2y (@-sbWRd GEhOH0IS뎁$@B+iij46mYfu`Ä,O`3‚Y/y '*yYs鬌r:O^W㖐tsK܊4<5' _fK4;v~Β^<5o/xjmAE6ǿs 2:9-ʾ96)ɖvuy (`Ms3w oR~T}vv\%4*unZ[h;w.>ubrݽyeRxF_,?yx4Y$ :L0&ki~t{+xX vIQpq(屃SG}W$5T*<5Lg1~# K!'QH$ kS<M8.<TI礓]\LqGa= Bu{#'^=`WEƏ^}.3*r*S…MH\9e}Ewޙ\u0r'Rg q?2*i`I1vΫ;OnfBIbg6EUY>m8Ck6٨omVK 7}z£s>?8we=n$~UǼ^b.yf{HvFwGQRTW75h FȬHVTTwulqޟS?×Cևk}-8)wlx8OL}(M53tlWD].⻏~l]?/."l+ ?mlg +d’/Vgݶ+wq rbHn?ݯ6V6qG_j٪e]}<vU/RٓIc?9#GZxtD+8"z2a.N2\7wډ k5f0f9H},PV9F7ws(' CN.J PRm^ 'Qno٩t~HgZ-0#1%irP#mG\j{}qS2(x߱Ha G7Z<].GLpB tAwa яkŝPq!hQ*Ԉ+"REN# Ykc!m((m,J;7ozmm8Ts mAtT7_K\>qXμDG6M|KL`Vda ʷqlM˚ ? ,e4~TɼTr`YosK?YKQ|gN"I"$SwPE$ i%F(&\~ 9g5\^_{~f}JŎm]?tj{{y{y{{y{y{y{yOCPk w[_l]yuDזgX|^luwo&\AΉD7?lEw-`4f<"~q2Q{Hv"p3~O02wofQqaQ%S"5Z "0O80sQc*tLv"FY$ӒL%*Ӟ M:t_XXY>j:epz^<|6ϛƾIPETKT"2Ng0|A:}՘;-jg:Ba\jHnU Y% xa.1jH N'c$t?ϒChC*Ϸ{ޏE!Yb??rb SePD` AZ>H aO*44%T(xЅj9":L-TMvZ~}zxŸ,$lT\d==4I{chtX[>6#ǜGU*D^S Bl~L1u[>w~yנ4c~w^ F!o2lRsheVy£)=mtSqT!J(;eN;h}"G.<Qn*\*ӌ X`P+)bx/ʙ \t02iEvX6,@-@[T9@>H,®ĢeWyw2 DEd/$hFσY њGNgá=QDz]tH];&<8K!_`n}G*|$oZg5s.P_kFS] nzsQc3(.mвOq5l{ RRv4'"|dgb)ҳhDcY:ud'a+IGX^gXH ̂vYZTM͋|;P5 l0O lb?Xzw9=1wX[>l:3usl5@O5;@r9kxa;ryCAgD-py=M&*FwM˸Wπ jh.-^-,i ;b)^j^3ܳ U.򱬝s16Ў\"PBť#!)G\oS}}#?fHfg/e.Ix4$&SM30&J%cr qDz\Lr^ t3]5HC(ASxvo=1D:- 'a"3 iY(&aR^5@F[❌3ŲvGjHZJ' nt e-M.ƳetlXe*> CLޫH@=r8o@I &hTL=n(|,˯Q_`T%҉Hp$i/D{4 #YX{2nWo/ɷ4RS"M+26Pѕ[8s=MR~Y}^5I.)Q,a{"$ANȹC!F7~Z>n:ѾV `m1΋ g᭶.9!| YPP!gJ{4|>:Nt#YkNJQ%<2⋴&2\"5B׸cymՔ6v_? Ȥ MSBՠ%Pb%v* 8!)vjq8'4$jX9Kl,G{h8Xn؄Gp[g1!lp~\!gc 'JqmWie%D# 5 F@R%AJJ%o*b|,'ڛV'dvpVQoإj8eMcrzK 3|B'R-FR4Z,f梮_|,S:N18'"1$hIIYɣQ݃kq]5Z>_ k`Ϩ7]뷮g7F%pbb81c{-0Ą]h|D! Np$H{KKxGKt`scYkWS/璕qŒl}ENF@Lɔ2r.QG\Nwe@~6d=qй>FI0Gv4xqMOe)tI8 i'Vy #55^w%/Z>$V˧7ϐ{pNz_cT =8`0^:!(OEO'RD|mef"2=1$:-mEIC=(o)JLkXR_?._)@UmpټQj:Fw,V2N7aԔm6~TNeK0rv_RCpd>7ݲzA[ 2/ob:'M3`*EQI7kh%|U$tIfH-j+( ^Q(={4\b ^퀢 L =f Fk/V lm1jl&rN'DSgOld}t q"ht}cY>am%yxx9MjPLTن5E-WIwAǟ:,7^/<ߘ7F={ Tp!5QpZeIBLsQg\50J ѲBδ݉JиbkFOudXEOᘸTϞ 2Na]AqZ55n0(wYô,o# E[^ yzhR'lfkp?ِ|0dS ϖ}E%}`/XqbM sZ e^q4"jXOAZOV$w~Z# zf0)(`tM OTK䙠;"(vSNQ>+f_/^0#a-+9jeI;eekz4[fل! Yy\x!{l "qrFF'@zq)n"*٣znzg>N N>۫a׶FGe?)Ӟ,b{PxD٣og>Z1a>!kEw(mZ0bFm)C“?@`Ҋ+rQe-ۘE2OfO|={5ᴭM;*hv1bLNE} Jc3a_r³5#pק 2 =D|{5VN3=3fԱtU~%EFkʿqbUכ&,$9[$ )i a!(gEup"htOB,CA ?[]]Q͹VY;mwGej ҢÜ[4b-=|&ݠ-)c[0FW]2/mp"t\H-k|vg96aVk&}caz&}Nǘ0ɞ=.Gd~yV ێL~*Pl0iM9. o0 gbf'[6+1V76i8O."56E|okܔ/g&oމ C# +P-b?aMa\KgHMQLph7siwF0jH ϺƇO!q+bδβLe+YY촯gi]0Dz\L8*/o.PP:&AS&[}u-qo-Qݎ4)u0V#WU%V]m6+>an@p=ndvMvATvܶcgf~E˖նd6KnĢaUd0.ՙ*ںF<}Đ+]dz | 2RZ+J^RhF-O " 9%yJiT+m#n#LYsȃH Ѥђ‰.]57<yU45 z3._85j -U4*W}7[ZOu9)kD(]1t* i>imZ!cd:'mbC9jL=G\IC5)LJN2+  Y8ic˞:8 vFwE~QE/K)6Oʣ֞X׶j5,D~*cN/xeE|6;[y!#S-/~=}Okes.6Yb|P & s,7" C~:z2LxLs#Yp/TMbbxMeMuYEa1Ɣ+"f(Bh@rsc3L=T񆮃EDy/?sm.t% O"SRLүax[H9_KaTRRPw.$Yl$Q:ExV؁+kcG~s+^y1jqNHS'*-u = ]} eه7 J2TbOΏ1aN $N2"T(#Z ))?Js{*^S+O\7v0f 3DƔ)iahe\>*-.%=ϋϿt,Rկ[OIo{:t'q _ܢ,>}䬂G~[&pvN23&:h[/)[9,MalZL4t^TkM\E uT|>" _$R]n(*ѫ,~7{F(SgywR( .o}ץi0Aj_C#tk7KC}5 Wib&O@vZv uTA6kn!a}: .[vDs+L1}h1EazCW@Z@HAA<Ӄgf&Y Ex)Ng`7N#A:ɹMIv4h $46gיާyۤ+0ZP{mPl@JvjA6+0`۰އVoV_&1fTA/u؇h+FTv*_|"=LB| bxpk^ei$TǧITkc[qw^#0C*p*N s(w+:a2u.6PuqqzEWûZ'.H>C~s K(fJWW+1|ZQ{U1mkS<ԝc; !,Whbh²vg@DT+N" D40aN`E b$/vQf~ꭄtڡYi E<8v_t$'+AKڜ؜Ae}Վ jKUPi0SVXÉ-bpt;A,XNZ f1-n 0#B2[PǓ\*"teۨ|. 4CZq KSj:pn!G !xEΒP>*s9j3,{AO iv冟X!n0Chw%<zU?U4SsH{SNz=ͭ݊p % Dn]:X?iy@/nSi;:@Wg\{໅u ouIU33ُ>0%^5\vv I!eC_n8%~诃+usٴGbխо:(SrC7~ڱ|1"c~y2@_v3G`w.ΌI4s3jM\$Soy4y^j udi־ϲ3NGwmTւr gc C6Tq崬@a~[/?¼܋ߞ&6(sd MbȂt(z4y9?_++`5ۮt F߿PˏP?[/BѾe}mV) D0hCE6#R Eˮimohʺ{訿hX6fVlB"JנaZnAC#ʾHRnue{ޟ??8 u[,ݸ6HuO 2M|_Yp _C%Ui8u44fiQ8ip0ScXn}!#ZෙU1Xulq u,En 9ц:xa7h";-"d6w*{G;C<bD[oᳱn0Ɛ) jGEP6b\BFf2 59 #&,O I݈Q2J]O]O 1vkqO8MVLfTk(bga/A˴4{'idDgzz|3=Mt`T](2p\Ԥ6i:<7<Nʚ5TC]9ʡ wec,JaKq)LRƅ:2v!ǘ;q:2u`+?a:9;i*`W] e`3If&:KLj-;#z68N82͞`.U?_Eո[J|Z'E×C'uӡiv9QzY֖,)r\Ix>0M VpaZmBNkpwPN^eٺz fZ$LfD YM9)mnuC>Dۜb;M7s`$*X2s_-/v/qP^WM}yg|Ua0 0 Y8icݒc*𪹤^ Bۤ Usٍ+&) u)Qq.⪫#F7ʢ/hl8ݸ v8/O{No,ȔK-1 \d$/;5Hid&C<$ۀφ6h2-ԏbY@rd ;~b il]0}%ܛ{m60NhxO~k!k&:ęmXHey2 ٔO wo}c׈"8n^UtTKV k))LڀZ&iΘ[+ZgB!5/wp7"6 <-(,8`؁%qIR,r [}%:`o!Fn(O6F@ǹrz[yT}h ߻6.ET2sn|a_9'R(͈L $20صa`P LxΒ&S3 ,}wpi-|Q1WEsY>`xhf%pJo(7׿a qʒDEjQYOS7 )* CW*~R`.x+-h%ۺ7?~5_"("AC/LR@6*/Y~s oΙ!% @vfpad#țp(7|{yn+nz7(O :.SdY]sR>8^Wr{{"U_Ҟ._,צ*k]XSV]R:#Uj>T0MS ))=O}H~6"mCUNY8s⎥薯ŬnH3vVIbE zX|#hKy,)t77l"5z-ST}?UBơķG;z@hC<<QQ&%d*cC)I(pS8nwbgDpw/2^ݵ#Wvv&)R$"̟Hu{m;Nz0Eɺ-%"bWWe~x/:12ˎL]04;!T˚'N G lj%Q[dCrk Ac81ˀ7Y{wp1vGIR`ۃswq _1D詞w68^)Ǟ;@QT\P@n+lBJ sIHDIL2}Z2R}OpL <p < F] /Rxe!4` 17:Q8(vD% JP"+}*cEz;6X3W N|`Ncc8D5<pfci3'|O"Y3^8c{DLFyh[Z#E _#S b[LT kS+5!hKsH[S7u>1N%K: )sKyDvL>' e,BEy3C1=R$BJ)728g;d-2ݙ.gqQ(vX >'q$=a/aI]ܯzs`t}s["duҘjqW_Vqڮu1n:~SE#%`̣ĵz1/y?;S$0e'=P>9KQ1=gU1:pź Mk~/-owDw^D|}XeRV vbfMQuвzj>;F-L&ۘTgS&)FNv2i;ylP4EK2 랧 `6-O[ς&. @ -C7:O%yb-l0ٓ>+1=]YbWb^/۟ʠ_~SXR$C˚s`˺}#ح_wIpZ^:R= 1;ݙ'UYkj+vz"55@28FjHCPE+rYH[͹t{v9ϢX߸/bkJԑ~=DU㯫2G:MӳMYeg{д$W0 _V󓆃 z6nvCL݇OjMEFSCcy,)(zYrk%UzMtk&uף.uׇNfc;] &}E1 h9RdIb`kL=,ͧ0!>P4nT:1m5V(2YvRx/Yvu·wl삖?ܩS?(*Da`390hqR8U yj&}p_u<_LݱS?#r\qrQt摥ySpc(MzMy]wmY8. ވO3՗EQzj=+]R.m>Ezծ\2jqqqᴜf=5?G&L8vHZzΰRMFϯgr_ [~;7v^2v ~3.+_+Ҷ0?[m>[W)^K KwUԽ>f Κ}$U~z8H|6٦L廕/7]3]f}=|;=+lGΎ0g(WzrMlyDjzr(1;J'^]̩y[5fhzk٩N x؏gl ZUGq)O-I$$,5Rae1$S+FQ&&cY.xJT2mߧb7-;Ny-҅+}g?MaĢz6Χ,l6(wW6jG ZLuhp&u-]"2FA //9l} n9<]o6K &ěϯꫢ'I†xp#jd `grȀݕ5M@Vg3w/Ʉ*~^ &ljyumwki< rY*^a+3hKPpzMpWEe2YV6wiSyI@z;8 O1ka+S"gaK޳Cfs~`fOc w(蘀ݓs$0 Ejb8'Vq8hQh6OۭE|0(0*Tţ!KAa Y %up)13DJ )RX<8}yąRh6J8dm$AoW^sgn8˳$Wյ99b%X1d!3)@?G2~#X 8p߸Fuynǐe8 o±H cJ3Fठ9c6N< |o̸O?z!>^{qAeKck11f (Wqt\ H, {MKpsgI7#0;<X?eHs0Q= 9`q|M.¿tOM967* tz\緗̚:8z0HTP^ x ̡'`j0HRi@{%<k<|0gW[y>+6Xy`^&AuS~$;t:fc  8ūxvxĔhn{H%fBSl81RipAJWrm$7 7p YÜ_ yT,!2kQg\Tܤ5?Tɪ_S}dUlqI\ +{Txk3-r1fPȂtSLH3.42lsĄ`KyĄ,,O{9MG]re9^!ۃ3SsoxEPD?2ԣ/.@{qI]?!>ܕYB0OLmЏAFח[dVchH!zAť7R Q?VP g 蹸t PE/sNorR!{LZl2 f62sL ff nD̾=w^~=}|'<`08Gxci@!46,/bx駋!>GP! LBK9s:&.Hl>u!=-ngbqy! 9$L:Do[ro#~fc#$ L@ 5q/a6Fq8z!4WM>0#IT).5t_NhpT2jw$ Eb_l!QyL/H\Sxfc8ڣW^keuU=<38]w]}(ir*BG&35txhJ' ^`QZӹv :}Q9}F¡r(Na2G<70c!G6$ƙHeVa22>puaWhjMU'{a>8j GM(;1D']oƿL;Y/Ӂ*UgE,yM-B;@/U1l[t _;,?z؋V ،AII"p+++*\+m Don%k>k]2OVetEWE xhvޓ Lvڕ5Oy~gz z؃_x_ 7/uz͔ &\IRL59 k9ɲfS`}>S`qR`w lx*GV˕]UUL ߐ6R܍ 4o!emz%xrѯϵz5)"裮̭Ă^]_ 'nW{hPYZX R3I lea誦}\eEdSU N/^Hf (, l#Ư # w_G7Nɫbh7G, qo Xm|=vOl jg畷mFxM^*/e<|JTs& jX]:n"7վj| Tpix`# ) `Mn}w^Z?~;ӈRĄTLoUy(d.ضG~/@UJ(ю=ưݱ~^-&}qċ@: ~Ջ:>| a5&xBdAc*(AGʾ~ƒ vIq/0Ǘ 4ݫW]'{=ǗPTQ:WD%iV&1_D{roy=91vVac-:o>abIWM`Cbͽ hq*f zSA.psq)&>1!m.@c=`673%1Zz>`)̥RI7wʚz#LZazƷlߪ8!!k 3 "C̀ёjG ֹV8 <<boQ$*jƸjGȦXv0 ,%`i>-v<>'0W.&FB(r4OBZO!|%T٩IAq5lJ8h-9V|x4#D% 4e&ԝ#LM$Un՝AG4}*cU`ީay!:R@ VIr^ã1Y zN}e%:6PJG\Z]c,/P]tlQr]ۭ%*\Dѥ?,6a$o;Y2jqqQEj4>YOMs}-w-3{bqe VѯUyiZ6-«CQ94g;kT Fجg~WoJM<bɸ( Z#Dc4E 煁VQcFfpLSq]tD6)}uz=\<3"{Smk/]v1&S3' uqa2 bDzM NpTT,~B2Zs!80SdK#38T Φ?~ZԄE3#WCfz7juL(*4q.L1alX_찈33"Z>N\73*McՌ̴Jծri=q5(pPc "`4̢5Cc"rߜ6P5hqRZT"` Akp]s )닩Q}BV7kʪijX+.4Zpb=~z N R6a1q?^G3gLێ~btn[%JsɊ0w`j" ˞}4: <ˊ f5J4n(7uN|ھaȐ`fJ6X[:~fPHJlP$3B/bEؖ$M/?67_[D lp ȟEcie"΁p҆ S `ˀQyrvF[-CG KRG34ݓ{#gGUȺ1m#5,ZD WDX 9[׏7[#t7"#Ԫ-[Q DceWV$ʡ?A#/8$CW3*r8ΨC9u.M2Dbk8@i@QJQ>%16V,1sjT29Dr aR9\DP~ƨB˴=&$4p!`#-gD,S$8c8 ݴ0/80LH*0Q5H nƒZ V!!$GE?>D*Y^Aʁ,7HZ$ݻZFgOϬ> +0C 9–P괇3F#E+[yaOqn oWW1yZ:b)f_:!!pӠ f/g4}p[ S1&eFfpXF}Q-_ X0_4 H ~0Cn/R3Tm$gIf>\CU4xis`;e0De:L?"ӶvI&(Џ3448}Fk*1\['5t8"@EVNS0q]Nx1NQe3@3~~FۤVA42ex$ӮAR9Na'^/Ӵ 0 c¦E~EL:R q= Z(3"y9QX9vN2% `&0iaڛML GLVc< q!bCbgU"#*jH$H3cmJw1>\ v<5"nFfpq(X\аuQ s㒰qy+. LI'E;S{e19g43\u73RbsL Ɨs*Fv*jz8̈jQk fXY1!"aSqK#38 ߌv\,SS^<UPƎ,zebmdFfp PEuy. _@o:X3YL HkDH"GⲹGt6e/ V6c&2F"E#829i42CX| Y 3kuJ 5uر520Z`b+t]3P̨~Tl 0o#|=#\c 9D}c@5M]ZCn!o NPfFuR E\-^4}΢f_/%/RU.Y&qS7F)1puaZ1uXei >ڃl?\^R?A#78:}O4u2Eҧמ3l쨖Y42CW}f0䨚 I )e;|Lf8OZzW blfD)|42# vf͌.Q QUt`σs"׆E*R&Dx HRa>V(k+|JXPߩ J2AOm]uBL6|-hyVž}lSmQfH(MA (#\8 ? "nid_r!]OT_ҭ~Bw$@lFT6ER" ;\3VD{&hVr{)gS.W /[#(9_nJWZ90 Ջi T՚y.UYgtj1A#78 ߻u;{d]i|DfhE$2lMuB50^SJdK73ʜDHQu=߷9%9 '٫ayfp GM=лY-[â>}6_WC1 Lֿrշo޽'r֌qzaJF8iHK΀ 3<\߅W lYY ,Ų\ӽ~ip~i`V`[*4ⳳS N}L዗}y|Ԛ~ie~2XͶ! ?dlzxiv]=.`UXb]m\gnޑew^;~UO*|㸘"c0g/k)p3r-ܫC= T gc I¸XmN25VTR۰+a5ZջongExcnG3q[nTݢa[ѷӍ![ yU mգA 7WW/@.h_}w[>v5l8\gv}Oko6}B._bhlexr{_ [Y]ѿ}:@>=!* 0jK^wc{8Wlnw~?HrA~Ę"eZ>w> ICRZ@3Gu=~U]]=v.SPGBIX'kY1xdѪBc# 3!?ʡ͗eqL`v[ h^ʢG(4t5L}LdpF/E@og/1Ɔ`vlnWϢ=-T0"tN  #Ρ`ljFގ?,ÙQ"n^B5R2¿V/o~serV2!-:фXi~}|0`Ry̢NsMΆ#~u,B%F܆kT8 DKQ*m,%l:c`Rb&V@8FҎY | \x]^lűx8?៊Pc lFt4.>X (X?dr0@-Dz`sUuMG;_[eXy3١m2' BDkqtd6Ե!+Crv;d t^u wXLtfq 0 )N,O+ +f܁׈!L N ,Py,*8FqDl2 >kPfvp *S{*02^9u{)4H_ ˅ZSgIHIz7rOe$n>hF.,Z8aB Ob4< Mh`Sl$&7Kq,k[ԝ<fo2(#ø{&} }=/5v@: ѝEgr?vsʌ<ې̓^Q=$Eh0ty EQ. vȠi V80xeԻRL]Lzpu/ u>W  .85iatH2\!r6poe>mX\,izo]vcJUVQ._ggMQR)l8ނ.ߟiqU$+.\YHɋlmNŸe7Wux0-͇.*Rm-Tƣoge4oIaFk멺顱'ƛ{n놴wpk7ufy Bh\'`bTnA`mrJJ^lmL`"4nĎ9l\dQmqgjW?5J0Ҩjl0gnps7_Wo߼y77=W`UqJ!$Cp뿟5w`ߢk]+ߧ_[/>-][R(ȾjL]Fqw>[`wYܕZY&0z3 ׳4s&QmsTZTܵ/U"q;h#ܪ.+ؕEu-i¦q@@3 22Q0J(#*8DQDHW] sj/g.ҭ9 &Q_XawZ͖gv]Bx/D!+Bba Ech kҮL'209};;5ckvkeY 9Iݛ'm{o䇭zZ$^[­CyC $5Uh'v׉>**$<"">GŦ|?/B~&0nIEV XPTa IO?#P$%ٟmEh./{Yy,[Lʹ<{Ls^^2; |I-pz58Po37>5?)g&:E *d0S~uZ ]u4[.UF/MyᅞzHХګwV%a*칤Yi4#}1+Yz1llOE)2eZrw|  J ~OSI'N^0ICg?t s-OOD\=%I6Я`Z~/Ϣɿ[P\ZvmZቪDkW9{>̼[ Wq =d9*cpžqGO(lqU lgA5UÌDFUha)%ޛ6HWRLYLY6j˔k߃io~Ӊ2DWOON'[&50|n$FV[L}{>ws_F#xݩOJp&Mńhٕ/C{n/#gptoK8 }%B)+^ovݘ:1g}:)biKkzU>ZV0Fq٧nlЂ l\}$Sw_x6۞3s3O ݋X1RBWU0/F&{W//_j#P71"rWzG^4##|MOR1G XR 1rbrVxm#hD}+Nw6?o); E>=^v y9'-O"7k|T/6RT/6ERT/6Ջ}qqNŦz^l&NbSؤ*RSb vgL?dPE9WY-ێ6yv/%' V.I]h:jyF4,tм5}֣'4qz{75-_uȉR4gKrTo+LPl2 Pt1f^&=qwu:['O>.D)0Չn~e~ff%Vau84۫Wf}zźSk&ɵ r B($<3@sVP@WaJ' rh:_q WՀ߽|w0W?ܝ9!⽪r&$ր$Lv%F܆kT8 DKQ*B`DPıHX@p) @+J Vioj1Rʃpp^j\x J~>Q Rv['+$\}n+ (X?dr0XIn9lxd7l"ϖMG;_[e=E&6 ZPC8ark D:L{Õ2::|m9fq 0 )NdNRafX4F%HD0 AP%@2AvJg1cPa4. TIXkºǣ=6^ WaGiJe9 rYލ\S} !'Gqa7(4K2ox5+<7UIN.H{@لI8i-L8o 7UXָ1;1.|y=̲ߪeP*FqL0F@{_~^z#ˢ6Dwe Yw+3OoC6zŷF=jI`XE"`,޽Ca#@$! ^Y.S3iûts Cϟ-UÀ -6`M@ZX.* WH ܛrzt&rPU`qwcミWj-s߆Ɣ|\Ug7G9HDJ*,-rW]˩M»[+2q檄FyivV7C T=|vv50]ŵ0UE[Wx-; hm=qSLl麭ލ­ԙ5$K^}b= !daqEJf3y W EOm}3fmꢓZ]W^tq1GrѹϏL#zH,?M: 9 Y0=03fޤGRp[*ftүבc~MvU)| ["N 8Z g\0}*Ĕ"\WbU9K0wX.:5 ɶYyЖB&O*)"s\Y@* | 欦 41#$+3ӋqK,F&nd69[g;kuMҶyG1tZ`} =3^]qI}>'4g$G[F5^.Fr:LlڂDWXti j$JwM'`~;k+&;*^/r{NzCՖd6>SE$g71`C7bU]b\F_ϊ_ _i Xֈnė7˿/plc^Wa|,y룢% ǒx Km%#S6$R251F`W$%h{Enzlt4f:w:)cML4TAMQ{:HB$qL{ɃIkTs u-P"wTD64RZr$+"^e%x[Kݡgo*nc'uxĀltF)L`jR >ᔤTځo8t%ךsuW wt}˰q>~S0SR7"x])_B]a^ ,iع'ZE` yUU>,Bڦ &np+$Ybs VF gWEqK*#JeF1PqiحҚdLHd*p.B!ɸ B>{-USEЛ.s"-L~WQS*Ca<%iHdőSHHHHHH8 q@B}\|G5JïQ ,)A^,A+jupwN68M< J.~#rͶq,u"0 #0 #0 #O!0[ ]3QsFM)E.`LXDYX[]$T^*yK>) swE='7|w~q2\9&Fvitɑ]٥]٥]٥]z7+Vaf>Xty9fT%{7Gڒ*$JxJa5TPH=% dMZZdʗ $o2:Dlx-\H 5Cc4;ml}cD |(lyN}jƛ ٚ6i. 0aţ*d$Ҫ#xD &3]ްt;}]J^#mOݓchK֚Sx%IY愹(ILB3+J"h.h$%H d)? yM%ePo9)HuH*PkXD\ =ZD=N}M,|3KdRnL(1b*g+Y8f.^UrϝTtI:}V|9.~mqX6w٠!%uu9[P@@8Jw9̘0C9{4 ſ\'GI2ji74xJ^nNi<~u>]8s#^ѧNjKׇSˋ{BA,Sog3rm㨪o.uBZ8|%ֶW4X [ߌki[!GZhؿZ(zl7M1k[eV\꺾 EY^0:3lCZftT}g2}6s 3d[NB;Mxf _˦E܈ 6^I.irnR[O o.>Ww%̰N\X5!.e//PW@lk"?D-k KajzudPW[լ*םL'_#G9["N 8Z g\0}*Ĕ"\WbU9K0wX.:5ɶYyWGO*)"s\Y@* | 欦 41#$+3ӋqK,F&nd6PS[g;kuYy)GŖ־t.}]` CrdH^SH^H^_fxCG_R*^q -iI,RpJ(1 |9˄=2a{zk׾ U]7 :acJYP)} x} xπoL|y^9?w}jy< V L"&NM2N eV1'XtG<1%bλg@7Oyw'بlevnZ;hWǺ- BgU*kuD+3Φ ªIQ^q y0k79ʬ~Ԑk:zcU1FD3aDG~DG~DG~D߻9Q5CwDqGwDqGwDqGGĄQJ#.B".B".B". ".z .B#3jKۉ>>>>|vgtA'eyJZDM#eI[Fm5$m*lN'ela𿉉*)1*N:HB$qL{Zԩ3Ga7fW}Vy `Zfs/~,FlNZ2;9S}Vm68>PTh4M ll4}bv5=V^=08O xY3+r:ՁF16oq1>\Z2nFgJc1e0LX xڇ Ⱦ Ⱦ Ⱦ-Lvz`^\+aG+_-/?HRpJC)Et8seWhe@osϏ{pL&&N)ÚD!X*qUA M6UR[e$1TU枩+1@# ,N⭅LI\"VTe?|tJŤLRL[''o`3򮻁N#9i1gI:*&nѧn5ΔP|4{!V=j9CXQJg6^7hIQkH=agY׾Q^e 4a2f7*mDƤq.He xL: %:XS$w*5̪?roWCH=5a, 6t\̶B Xg|ꂙmZYW,=^htyI<3T~F^ /ѸxA1b6ڪ'/i}PW2ɃwmD`Sq-$(ӅYhR?kcPګIHC W]qkkh㋾[%=<7n\ѠU.RyG,XحV޹lSi#1d/_wԂŬnc4v!2z҆]OO%xޤZnuha>/Y6"eGrX6@ֵlF[;\Ywh%AU9\m:Q;7վk)؞Mdf?2߸zJdCUVJdzíu;;^0r?N?PKkOӲ߫ZM6q'q'q'q'LGuL",%:]KV!,%R",%Tm\*K'뚗Ix^Pu@䄷dR8RXIe!@==]`14PP PksoɩsE=$0=3 v@S?h Wo='є#zĻ^u^u^u^u^AeL]@>"ރfXT H{6nEm[zh;3K-DIN?^ZZjȝgudE2@邧ts# GMת/t7Fh3,QPJҌ ]wn˕Y.ޅM椬@ M'SsoLg϶=g;qsУ۩iVٳƋܹ{4uKo3vvP?I.Ϋ?;S;ڿwN] nlzvθ9]݀nȕxOw6܄筷Ws|9XsE]kcAzp^ʿw+Y>m.&O)qsE !E*)!*DP1DAM02/BPg)88'k  bMq9.ZK,Hy JR>[SGNjui;+b1@|SEcQhT>ʨTkYۗ_棛cڕ?`_,Bp^<)OR>)H$S" \#@ƕ aܓRO^?ǎ=' ɚ*3擽&ȔIDi*-R L/<"ЭgEǣʫUӻ=yg0WvA*2HS(JEj;qhx z6DMeءI L|QD"'?x~GiЀh¡C狙&%+nlNX*h)f8ɻͷ s9D]ŅM^>Ɋ5jp\+|XqpH%2F:aLΓ/e]9^M+ѕL.䃾H=8qMn2'⬫ W7-:yzCYE(Mj` LeHc4uMׁϩzGmh+Fǣ\=5!j =(?=V5uLGﴞxTY=lWO\xD=U.- U{ԥOȷ#:ܽʧ4o݀=b2 SN^2_h.Sz\ {)u]q Q`v1ѹ"%R(˿4Oe&MF x@0`Mh@0 }?^^ a',!0_#!q![t]_px|z]Bl3-.Z{Yr-dr(%s,ϾuNG;!!/7'$\9'" >'2Y@ Wֶ|ZM#씣ͦ b'~wJROiq'LvyxelemL9T9vw yvPf}|:H`0:()Y| Z #ZL~ĽKj[V`K]]JTlW"']ilqO` dzx` Ӷ`wy@ȧ\NT6w.=> CӰt ɧf20DOC9:,biu"v=!MY}-mt@D ]UU'4uЫtB{=#G4wC_Fۀ<)F | KT`]ւe6g!KDՃ_vY PMA 60K`F4H oFRkrO[(ҴPH. WYN*3\`YdB$$r\k F@: ԜcD91*@_6èP\/iuxg?4YIlTsn,ZՆdo]:t>NmKx&:TN!2΁EALcLT%4y!TiA&Qx[Yiga6>*uC4׊X-Vؕ[X|iR,x%ݶc%b2g3 fZ*hFH*+]AL $EJr ӌf2-EYNPƍ#$Cuv8-c=kRAOJ]hf #JarLs(gLk,jBgҚtQXV2ʫ73J /}˖1-za,ZI I# )kۺ-j'߶H~ץ%m ? GUWL3?{mrautp&T}s=i;:\_ʟ=g:,=OWb٦aҕ %fhv vh=G9d^eͽ Xki.еBcRIH $Щ҈݂DHE|H(==clu6 owN6,>Ꝑ9Dj|Gh#AA4La ,qC)GghR;N))b݈&`5*c BjEnGp&dD(aL Q!bZ#-gDsO( QiJyjVߔn5?+{MSw+wv`tuVCՓ;.@69E`T3\`$eʙbL ZEx:sB  bٍ9k+n?Hk2f,:Y^qfW)V-޴E ! ͭ[ !a0u p;T=CYƽ2Yf)QX+ڄ]oLT[)4X(-rsZc^h (PJgDښ(A@21)2Ϣ0VBi Lr+9<rj e(H R]ugF5q{Π%,#X.D\Z%#4ҚTY{S@qJlQ-=/3TE5NnUn^'7"\ygg[ڐɖ|Jљ˫?]֝w5+K힜W청ߔMz$a}UXU~ ;,H}lń}GܚNh՚8Tlbp֐6(y! f#L0YQ`--2%Atmԋq8ogc~@zDb娕ߙ⡇:H[ǚK ]Cm+ RIK@2$&Ӱ@ CVu&fzwH0:4O@0[Z[cvfEjR95__[P D#щrmbK-vU_z%ފPl29?X<݄Bwx81 DbZP$60 L&<7FGk Y((v<e ر WCc3P=wj-nR!7Ӝ>K]ЭQmt{D݋GT@m;ҹ=:wS!ņ`:o!!I[m%UTٞHσԜ<ˬi.@SB EDDѴf5x9?FI: ̹}u5R'g&($G]q07qSq%Y*.RWS:GlVͱvvWf~58*p/mjĞCqvl@[EcM3 Brl-Q(&mF&Ħ͈@eH"[΃g.vEޯ >Fz)0j! sA`99,&vb}0QD0ˣFnMeBd&!?iԓP1J{m T2`Į{%_uqXhd>8țN^3GW/Ah7I5Hu,:S#mMiaqbђFjOta/ywR,6]0LȖRvr͡setp ;BPۚE$BiCFȇ X ?V ۆv49VrZcc##жΎeXj?IH&5]* {ѱcz7 @~/ٜ2=CAb.1l;m"&2Uފ(;n_M+3=K'&ex$F/+{-=K1 Oei;&+xeg,)T%>~& ~/fKz?GSqL[gNE#s` b7{{| "=,>k}W;ca~4yK|dS6g}T*+rLfȭKr: 8eByYxM?ۇ#߻W}!pBS4PliE-^n}0N^ZεB@c^k?V݋r*w| Q8 zb9(8㭬hǼ퓓xC $ډ -jb˱nsBEc蝈'm#>V59ʭns̀SŕKZ=\4W|wg/ֿfO .u  ԃ[=}CjbO$ܪqxh5U8s2m-^yhm Z;E(_Anyq?iI @mn}GBm`tth[ iPIiгU|-mn8wʃ6&υ[L4O3pah4 !o@6@]Ӥ^dgEBЉ>"*H`7cZSx{ϵj*M&mvmEt5ۃz `Xr P%:n-GַdzeСT =sf2n}N6]Xm#K|}g"nn>?^ghGL@6MOT˫f:Ԩè߽Z;k^̦{r#J]堎nouXlcolf&"Bkt/0cqe#ʩcAJδD11]6 T&BYN,X[S7懿n17~&ҩm NW{zXi=^wu.Qph(B 5h|'ph6D._.' @ YEXK:;[>j:p_o _~zǺ[ ؄ 5z3p esºi]BĽ!"9"FƭFxI.-QR%c[B1\"qZ{VWoXbPyAU#ih&]?Pʵ2 :8i^Arx]t i'噇0Zg[x;̬O]fɋuA@ŦAyԆ`LeMbT#'Pu%cvE&somӗ*1iAFgͨ}*~7p6^ԩ _Ⲻ֜NU~Yuɯ;6Aa(U#JFzSFcv:1"xyOxOOt{u%/%/%'tҟϾ kU1HTc41`'o3Hh!KY#Tb^k=rz6>w7E ~k+%l/=B3rEOo'f=SAz|nҳeM3'g^Fm3zq͸5 2K`NLD7w'cw7&&{{e"".2Mqy w{ը%o[0'#\˄B-.͟ȁ\-u컜.yy 63:*!O _'er*O4Ji5̋6[#ss~5UEҸPm\uFʈV>1~yAo*^Z䗟 l?tw9o][[oKtKM zO`axpi E4bZS_tt{VރY= OO:\+^?]RzIQkżv<#a1l!=,ڵT_ aa.L٤ۭ9碆5t2X է8 ?t y(S \sݿyaG~56L?Lgolq(wa쥭wNe_H6ZY T2[QS:cϳΡju N_T\P,F" ~;䬣'tc+DA5@(& 뙱1$A88'0S#FMpNTIY:B +P9tsrHwԑ ǧ?{~?gzJIb^wW L`xg]{Rrsnnɺ&0cWO'Z<uf ZVM%&lv`.Ekɧ^s44tVQC{u VDie z"(!p$TQ),-qD3GMi ~P]\Gv8_X7{)޹=pGgz _z~נ߃fA1|> =WϬXRe.\K̜hyܠ37.QtzՄQclC Sb+1i*ie8_z֝Zx씅J-cQ q OC啡Űp>-8)͙]^œ-'†K.~=sgRz|FǙUےPShTi[Z$۽ `#̪_7UrhYw"]a>"bԅJ msġD!g_NKԒƘ\KBxa \@+uIJ7G~ G--|.6:+8L0:G =5{GЮ*p܇~0QJ˽W],4hrZ$ ~Fn64ID1Y2)c0؁b+53B6i5hm賅9 l?lzu5 Qsl劻?>5:[eǔ7FzxV:pLZ;#1n}0R *if[М>(omWRϽ!w ~VϾɠݪ,370Hq64ظlVӇ0} ptD\ KP$k91 ZZ % qP*q׺>i(w2I1H"cCB9lSsud9֛mB߭jt%}N]>0U{05Xi2CVx-.Pl} xܧ||j7ˏ*"yFky|;QtqKJ 0[;Xoȭ\;׋7G'osHU6FTc sp M8a)5rT{_|tuD#VHd$'h$L09NI} 9>h^r>wqp6b*0qA4jJZ@GܿZ r5 V)ΡuHPЭBd 6* Ν%Dp@ bz@MM^J0O*=%SHR Fp)$J| %gJIH Q`nH0ѡ}(` 3OÁD 0W3⑅ȼ+<| %4mu#.M ̴UJ0/-m(!!Tib<'2*1Pypi"(qqȑDCP(Ty8ZdSA6:u[W8| %7XyFvj*(LWNYgh=(`^bJ<P^&[iu9Nc=(`^IJ3aQqc3^-|{P(¼!gsP01)sqP(梴?ϹL( `0ga(5XAt" QZ\#RLiIMV F;} %g |`,ȁfs;| E4*p9 Un59_ .Ƅt0߃B +>N F&cQfIAQ1Py0vOJ3|JIaЁP(ߋ^Kq~'ƕ(K?4) (WR p-ܠHS6 '>}gͩZUP'ܼ{?>CR"IO .*EH9 ib^yw-g`JQ^!^wE]abv=*jYČ77L>~C.9ն=_Ū秴RN6(}>$CH(;c&Grcmp ex*@f7A|MƧӚxɠK'av|WIZCZI (0V aI#0"!G4CQn@~ KxS :Dcv^A>L=*W.H*Wʖ\x& OBZi'jkHy>!g iH 2KoIsl⊀\E."Eĺ'#I67DdFV&@x Zyi u!ҧը:~w0zo?µ`8?0jؿE~kRbDW* j:ןߎ>Io7[`T{8ݝ&#FpMl9]r$mi{‘#Px\?}F)^N(32G{8ll3Tsk M:h,XX-7EWUp2x¼[1/8v.>sHFb ώ>HllּƓ0&]ihzNBu?{tqu`vb@s8 N9o?Zh49r:J"[1wHf7_;g:FOD46丳OźOgKFO<<:DsJ~K"f,֜\HXy:qU/x4_WB0!`Q:/dO?ֿ߾|?}/{^KqBaU!w]?[bW~S-S+v% ׶$gޫc3{V[n@JrI~}8M5=ղ_թ)B&yćpUs=ɣZ6Ѫ,`]^7\VrgB|Wb+;ҏq7~_L+.`qtM.P<&ZI Q"AHŃAD0T L)EQeG1xsuaiX>_cTֻPGfpW8:bPBP(L k dV`̢Ȕ}C&:S~R|{3ohh& |pAA|ǍmErE^Kz8QM=DtMAW2&qSrLf\gUbUhxፖn -ǭi`GY hXL 8ʔ# ^h}IX%.uAJa1֤*2m ~-'~M-{4JǨ͙Xjr B |ZZo`€L`uLk6/WCf+n LߔK[6u97^* yM<̜d"H\!# JÛTBA&7Єe)l-slI^1d* ЂxoutUgSUTe5DIaI`5 .mH`2BQ9 q^19@DZon E:TL Hh\dn&~mڌ!vUk񋂴'_!]O"L$"9\QTD ~Jl*'Hřu]5odGA4ɣѮwfE(歃 DO002HYAB-mj-`9ΠLl5kH9+k0{έl]0p!]-r-@"J\E2B ]4, mAV{3%Hɘ&Ђxlo> % 2UR`Ed>:n9HKnBXZ4J3V XkH^Y: `Cq!# ^+͉OL2HNyalmxTj-7R'Ť3 XwddĘD˦z8CFhA ^RCVPpLF(t eЂZ5 Q45d" !#4 ^'$OIj'N 8ڡ}z2B eUe<q `1„6M9m!# ^ۺZ4HvyR[KZ %]t~-'S`əwLZ ]6?`&敍s<&f}QвHeF!2B j^ʽ`@pV 7,`ɌC,:x9xǵ9#^yQk'δGu*^E(#):LD&QGRt%a _W]C$#˒/(S!LX\ 7g 9"pڙ b0&H"RΫmv|Lt'+!(+xDɸchwё d&9`Jm)g[15DBH4v?r/}OlRS fXä] OȈ1C~a 'K2(jMmB0N#-46S& SxؒJ?GJ_W{R_O |uE"6QA* u TґNHdϑ>{ ytlV'KLcIܻZP,06;v!/'_-* EdaX"i':.mqMS޻L4Z6z)ۅ#!8eBauMUCLYh* !籶K&6M"g+$.d[r( NzdUR[ҐJIԮXI@j0l9f&?w0r-z*'djZ-/}qo=ͯ|:O姯mKeQioR!`sr2]џ7V'yŅb _oNy(6sY_Pӕ"ٿ^_Pdg Ixp|oz}Syo[ڿ!qtu -"Cd>4ptj  ( dQGϊ2&Ѣ9:*`=6$=ܞ)mqb^`%NB(Š Ϊw=mr{J3HYB[P.V8Ɗ{{:,>g?tֱ݀~ ZQVPVB!G;hw?n:oaPdY xᇓ&  T3("xMGA}PyБ>v/:rap<ݒ^6܍FI{c|8l@gSh>,Q*Y26>Bh;_$xoJ“u-l%pgv3Ұ ):-`*m&U+Ȍ<%C{-p<\t扐-ZhؕNv:t;Qs*jPcb4*t)`URBz;MiNSӞ݃RԑHc{9t&jW$Bڥhq\{?wxw't*dRoZ$eü;$fdw [%vi}zg}\YG{mt:s`:b!cc!FҢXJo(UYDYA&4mΜ=Qخɹ9?t6(}>$N%ŤGs2Ժ #FfwuNjN+5_kS5e"o |A+2>{TK J)䕲%k6{UKAz6I;iV[Er@eҡL] %,%Zn%ͱ+rYg-HbSt.3bk54n@ʛD q"y2xPjz# rFu W_]j=tY QŐ>ըF~Yśjz׸/ {F-/!LLf16d_VnVKݲdb[j,RUWd+F0?ToRh7x+fcʌ<|CJF˛Av(bzkd*aY3HSPݠK)(ԝ vxwVNa> C]ϚJ`hUtb ք  -~%qjsvty4(*+ŧ>6>V?ލ&Dܼf7o[؃DJ*ay*M6:c]/N\SFJҽY/?fOWN7Ηa 6 Fq8-Ջ]$ƣ 9|pa3(Z[K\斮ۚ!(̲|a4'4iw@6׃)*A[m:j[_r1e:a9ϯKUߖ74mgGW)qT*@'3l0=O0];ۋ>^`.u~[Xq~d!Ꮷ ;Kb{4h|ve]niHh/Q#ˠt)!N~ҁbxWG盏`וN{KfTK"sܵE>r>*Um{I\˕=*xY6>YʡW$!i ΰ)de"2J(#*8Di 2ZQv{| zC<(8RS7(q(@,C/ k(b%HCNu0y5=t3Xh@{o3`3x:?ra{NtU"NטiT(2u:4mU9~8Ks~83Q+X+.(`Fdb>DNVi>a^DR;1-K^ wE%KJ*)躡kQWCm2NNni]fu$%9S#ǥy*'L["CNN-<~YòT|17?w/1@+aR/1",Dj1тFQ4(2&"ibvf`L_whQ>4:+];_լR\rwzu<ec&E3hw*j KIw0P'x1^긷vю[o;$Nq2Mzާ:]NWf>#>:N? Kw:g[z8GTu&\/;WXg!Vʱ$0TgUń6Vd(tŬa8GJev\kYU% (p_WD{ԼTr>$EXpW̳|Gŵyp9z%VwM(bsz}fPZ=#yX 6?]V_NcAhX,Q` e +gI1ЖG98ڳǹhۂ 52 #YD,JԱ#đ!q9Ri,ns}Bz%)w 5hꙹu/@0 \{7߮33`ldbёX#)9L0+a I{aE$"_hY"4ag:";Áms 0uSn{w_-1Ł{(V$P]"մ WDEα˼2(3޺SԤ4IIT`*--I.IF_E_5kh\/

y?i5왢G#EU~T/$ė{t։S;n .LLȉB/~xq;\XtɥZIL`(kJ VL" H{43$)'0Y`fbe>%R.xRzE!ׯ|9 gޕzKq|dyIeLHiO3i=ʹ4ӞftT#p= )ӞfL{iO3i=͐)q-x=*ӞfL{iO3i=ʹ l@ڸ+ﭽ_'"vQD.#2eЗr`ۉ pPL:)Б% cs`1(_0m9C!,΃9_IX[" zyvFT94BvzU39s {\rmLth@1Ra2Si# RbB(KaVZӘG"h@QAsgl#R*ܙ8-x#zv<%hbWKSTN"{G,LvYqwy#;#C 9\X$}1xTg=GN ;G0YX"!8Ypipa SiS`B<UXA Ā.&aJY9j{lV;vm{E.xJnׇ'ϡ;Cۜ' >x9V"3bǣvF!rQ"܆VPId(D&p%,K0iv ;g๞mTLHO \Kku/1%j/G*-. eFs }ȁE${A#%Z;FhʶӀ,ަ\D, M9$ޖ)ڄڰ6*HRHNc*̰4N+ !ʓ$Bx=+Ž $`#5}?劤`Lu*bdr ;NS*4Hci/F.ks-MBk$<1q3B`9! @s1ʂdC֐ ۇڊ2Av&4Qr(9T囷9 {ot-*k紫t1.ߛoa{oߔr SeuKܰ4VOͼޕy&?^J7=# !M{goC?}9i醙eΰG mLA}?tv.Swf2\Y9 , uQ?k~+aUiC(XL-*+$`ړ]^$ ici>g`UtwcI*Q}m?7Y͛V tU l5u)ƳU&-W|F&?VT͊-FxYm1{zvzUq 60]ŵg0ll^&Qo4}VA; AZ⢵%7t ioFf#q5?QLx4k\9Lun.mcmod[m}eŐ`!46>'.U}[8Ҵg7<=^- Љ*f{npsLGy._||V`m",YcEi'p?ĦiM+eߧ]e[ڽo>/ڭف/Q#ˠt)9N~ҁbxWG盏`וN{\]E%կRi\*w)DO #:}U۞tUb&:&~p' 1@N[H@p]L!-QBQy'bDHW ܻ ͜WTwN< ^њ(AAGb*4|XXCs-qG6(trsRT#06 ?ra{'ʭKkNKpU_:NRy/E͖ZM^{-TH`c)qbD2K6k:z"zz_Kfu29`}⧃I  Fdـ0fl9h N 0$MF&J+.og]iXRY @)lp,*P4R%15cB{QY[B"u:+1 \ P$aAq"QFI{+f>lбcN:M4mvP{'T(OP*a7]qB!EdWG]) -A)] M1F"cCB9) _xU˂4ه9Y ?!LMĒ\p3&H e0"N.FLjؽ@ղ8 '0.%=l;u ϑ ޲6yQCia\cP`0'zohZ7 w2-\?|6L.fnfS_RChtNh\*'gA+j"$z~QH:?Saxz; nmqH~VV! #Lv/:)}H34~:I?z,؀ vPCyqG}r*xt?የun5ƟYSw=g0 x4T%چLFB.,F"5CbuqV|TbBʘNB81x-@x=qh})+J ;#Z3PY;i4c94A:pf};w$ɮn8`hD@R#X"p$ҞR$Up@/~h!^c< D͕WRJX$Z Bi-yƯ$0[R:L{NĈv!TUT&Bh5$X[)]l8[3fΛx-z^h^C9o#[q-gexf\4gӟ5ݶyKGm{uI<7/雳-}sc%,f*s dDHӠL");;=ZwΓ'~G,5L)l#@7hd"2q[.XK˭4$K%%>ST -e]ٔ5nPCBԛon09#zTJ׃exR',BClhbLMV/ 9#!.%w"rZ#α@.jk#è;,yk]2GZgVmȼV%MO SxRPGb'TTy`IUrG@l5>utW%N;B`}?pqsёgBg I}@e4EB?JKpcL.^XC0&&A]eat4KeFRy*qI(g̨`/jp{ UH-$ ni(:[𑇳FKadNPO.W~8z[<Ĕ zq9.wzF~p_lo{#q,(Қ%W9m.с7 6O.˳:F`& 4]^fX٣KX*.$)Ïn{o/dYWHϠZǛ5ͬ7_Uvp4v?lo.&?N)*̂-ӛxb҇UL8 O/>l/߾Dhgfw>۩;,>Fiő$F.inO$)\yi jӞL[gP62: HL\\,%X 8Zc;,J'JqKh2 @H b6`ɒ7z ɻoPsa;Vsaq +B@e^s @ f ce1hN:&1ʸd2zs@Efq]P"I%ܾBnS)zTII /`lQ 78"9)RF[;-$A/#B"ug}aegg Z˟>I#&<2l]0 "@,`"=ƁyS`O\Țh[e"<~Ļ,eï uT7pHUqj,;Ȓ'썌k 5* $8!%E#Bqq[ͤ "`LD$ah$[}P\Iΐ k"]wT| lxtb$ /*xd<ásRׇ>/?]cH󱟕f8ROF*^%m >yQx5rZTћ6P;=ƨ40hظ dd:W5OſO6UkB%#/z}z4{v"F}ߦj6Br ?[Ӏ^Wc'ðd{2FmvOc3|F TAjů*bۣ/%&d%yAd}O:CJv[2kf$BP-*f-vp:~3귩_%wakSԫ$LZ^>.E-YT_U !9i:$F(hR 8l :%nU Hf &9_bW|-_]`7RLok oV'S3.n''~:/~`t?Y0 mXmpiTq*xS W1Eʶ:4Qe ) .DqCSY '!(fB6`d1t pc[@aIHeӂR4z)D%L{AoX#D{E3C5 $( X@ry QFI{ TC6J`-[VVMhm͓v*jE8u:cE0<`.(t+mHЅg?\A6F0 <ޕRX~>muǛIi8}{/v>l03?GmZhH \Ll':ǝ䊩(HLQHUR&yD:υfB)jcԊ 'ƃjǹ2D&Bk24C޴,6|ΉRskHAڕT_^k|o!!9FZ̿@㻺Ceĸ'ΌX/AsȔ<GDp2Y-b(nj搒`ko4~t uh2L\ϔQ(,$m@x -!J.2p=^ g T>Uu vwAo2M0D \ xT!'dpXo/ǾJsI' 98i ?S.` X`,*"B(O8 `C@* dhFbK_q z /+J-)AX䛚xt,W#j{n}l)_e8_SaKqWYvp4@nԎ~`ʭ bp^ NB2Z481#CDI)m]3Tgqdŵo-IRqzB1> Ġұgrny9;>9]@J}bC 3~s)Ei~::: uZa~4}`8m\8ͻ+5?> =Ӣ`<U_ .f7ޯ;_ř>t>\횓h0ug ߣ&wBAIޞrz:wio7Ax'<يOƣDW}φn~J^C 60%Fr篓Ϸh`Ym:̷ =MTLk~x~y(::/~??;݇ǔ~whq~uROznu=5Įe\ּS/&!m(q" v෋/`en'o8]Yy&#8lYRλ=ctݨzsTF? SވiWztP Ixw~Žly{$q!H@}2I.pƁ胗)%&h/rPey4KIi2\CTf-hY4;6VCť'vghS唷9JżQ_ ;_Srg8]E֏$:lM rn?}&i]p6XM/0ǃ@$SB }&eF2 ,Z#g<&bU}VK [?ϖ=, ޮb6 #«͢{jy_e6rp;YY;`XYWpxfPܛ c놩GjJlJ7 LjiV76a]1ms1"F<=-?-;eҚ΃(N*z,Z҆a_덪U8l'|oRgR٨ޜGLj 8 Rĺ!2nXH\y)cb[S<;|>q*&LV"SNM{>ލB>v;DI4FūejYZūejYZ{ūkYZ-Wղx,^-Wղx,^-Wղx,^-Wղx,^-Wղx,^-W/1(f~KU%;mN|q%ɒk4f69tChD[Pp=8QoIfdKŀ\ Z&DJÅE&9C)ӆ8-"=XE=˸r:sQ8߽H +R$&#|^A*#^%m_ x5`wOz7y}oZ|XV ~[;Z^%ADQ\l`[d8npɃ0I,O ,tDMHOL3Mb9̩BUU`-X/zGz u lD$ PRjd1XC#wieDL(q䊆P)1Zj2uoK+(ޟ"&=CLlσL{7׮KŴsXJvD[(:vv1ˇ_ sjaG]6itZq4&&݀ѶN8bұV~ p>]k:p8kO I⑓Gem rUU"MIN5 $( X@ry QFI{ T>>X"""mH{(6kTԃrhz7/`[p WdS<< I5߃a%rs\#X 㘲16&^)xjū.HOڂv#u⊟RO g%&ji.nW$с>ipTȬӮFՖWZ[/tB\ #LmeZ;bZ 1eÛo4B.'Tm)+"@Bu@D㢶 0ZD)lhm\[6 _%𿎉 )JLC"\x"e OJT6Nj[4g촺^q5{*v 3;a^=:S "rY.ogmdF+[EƱ#-3;C- !Af6= I&(Z5E'1J@fH +*Iz.&NOJ:L- F/cJ+2 G Bȣ1vvߕ"8#'٢dC"bk4\_O'/2=_LC;Z#2rv"qg@ 92qzs\^7aH.yZ'I)(dL˥6QN3#P c#GB ROh0YFuJ+,0ˬ볳W&'whƩ; h;Bi(p5o8XCt >>>>Uʢb1Xa)D"*h4Бh $"Our!(\Xύ$!\vۘT&HH8gB'bT9/oailvB>}d G%|9l}"bGr;m+ N|3O ,^1&f6NlV)eH@0AdkFhX#UDI$RiMD1&bޮR!ʠiԠRCĽJqݧ++)Ֆ̇'m#9 %YdG`d<ĻȇS$MRZksCQRej]U]WW1ԭH rH# be}0ňhA #(p*92&"ݱtZ [x) xw:qﶚ} `;k_Zv;駲B=^!'F+g>uL5=6"&PF%R8U4*k:&PfTetۨR`p')lr p8MVg-}>M7j^(/5'8%Gyp+7)/\r9++J*z^/r%ѭH3osA#S\]*艬4j[q?7\7N;z 4оM`VaUT8}ሳZ\: tPP]olǻ{!1~|}=`[xS:m@7g2H%ʆ1+!qydަ &(xXŒ9@|ə8Hndc$U:d{V1P v{aA+4y¼cH*fA^,Bp`-1M$Ne}<(KuY7KgRcvK)ܾR'qjiv$%osשjrӀ.&ٱ sh]Zer&[:l,:9vyx{KôE -7C2l5Wl[]"Ϫ-Wvr+4u2lCvzUyǴK:K'/UG;2qq(UDx"S|1"?絿cꡏ?<-UI;ϳ֓xw u4D{R ,z̝vX-"ro0CJf/VcдZ%2zT } ,k7 >eAFs~Lu'HD=J6 IyrQD9IǑF^pZ/|[ 9o|\JErHx2M *6R~H1 wj﷡ZhC_޲A_m cسkd=<>{H{Wulwm c߻ kx [J.Ir(9/*-00R^QK پ#RA)H:w.MgE6Tf_o9_b<d/< U 5]|pw ro;;ᅒh-Jw @Gm7 ,3.Nr%u- $ki|2nI rMN<,fhd0$OPP0+D{ "EJGt`ZD"rRuV8/"88f `a r#ȁL1]P=؉N4hL+ \J g($^2!YobA% 9g )W/s)Ȥ k01(Di'J;$=k ;QtXYYŁM>Ső-<ȟ YwpS)D yv̀ a9? !za yzz4`eB[ QH;9`}(2ml$EXKQL--FCSp:f,qU#dd\"[f.ft㮟1z~gZLCbRIU.24EOfa'?gVEcg!|^3{ D(5IIV$0Cg#tQ J1ugf2 ׶w6> }W@0! ) J+$MUa|Je,>gVJw9e#S0Em뿪gգA"F% Y4u)@/7pN|J^\*LUlLgaGś`f2 9Z\A>~*ז.pʹy 7[~DkƑ_?u0yYD@>  ?)>ꅞ]SqT6Z4jL`9_I0FñK]LB6Yb g?uJ0Щ7(_g7`GW.~.~՛w0QWz8q+0-6MaBv?c^`|{ CSZ hPW f\% 'l͎܂(xM\ 6 Or8Oy" _i_yp{W꯫ҢR]Pifp@E}M>eb~,M<̗l$!A7a5S`DdPFTpމ#kF'%鱭 5 )Z:O< 8RS7((@(@,C ch8A'N'NIތ pbtu{xکB>U|%YQ*rc )i\u_:NbʃlYFv`VoO c^R.eD}'-$` R;Ē逰 OjRz\TknAXeOY ؀v/ 5`qe9o]JI:YE-#if:3g a^fl-aC䖸` Xw[!/&`hk>,(?]"@$JQ,ͭ$G~[e$YS >V% ؕU`Apx:(p AB–sd NosL(QmGs+:2,"1`rX+I ;2/t[VS*cZ#H'Ko gwv~ `!?Oq$>EEjf뗱` 4&flw$p*e?r ʮm/`$?:0Mw. pa6ϖ J\gTץ,s@ohUf5 vӏS}[BG$PROqѷ6d~թf<~d?}_kiev >}~}tyw{ZL?m>5Ք&:zbBq)aXtSidkm#WEtLb<,0qxu4QdDzI߷(lYVKm GMQUuW1S,0΂պ{CErڡ_ :jsFxI.Rt&Ϗ* d*yb kp8 mv5| q>mtm)u~E >) 5p?$ x}N\sqD%RGDk9ЉL&!"'H<'JsQB2F!.{jiy9/z/g>,k| cv)rtяٕܺL͵@hy/z|;?G78z~yt}ztc҇1;n4G.OMf~{6+#uϷ=x}|NOԳ|ĭ-9]o@ؚ_[yZ<sHoK~U8Hy9\m(7?.ߍϛέW$'#rNl4Pm}:$I߳~b.t<֟$ &PZh.P፧:$ָ-yV$fMZPFWedvx@]}Y-pHzv,TW!JEGBWh^7B\ަ4 ʊyjպᜈEm'ZD(TO&=T&=t&=XfnS$):&\IQDyS*B`AKE ƿqb%t'ȤX2wUJ0l<*7H;NJJ#"\;AZga OEVIo+@ G;zʆ.Ug3'gFqVH R LqH3 `yL6<B#&.I*`5=DӓHP,%ΘQ0fP ?J-XeGC | >*<TG굶؜qĖ-^mge&1((iLLFF;Lt<}>|C"֒^Г HĔh/fzq m C?Am*`(?~LWWBWg<|A3kHLg?ׯ>?!~8"*Ǘ^#MsUt}R.7q(=Os 87k7IIRRI,ل:D9 "Vf`^ *t%; s)5RozKQD@5K BA唷HKRA:Yp!Gٸ4!yX24YL34P{PFB*sf ę‚$Lc>piEj$J#B_/#}C7}_C74tsz9t'dY[4^q&ًMmba4uf>~yW!ģ^O6rLƮ_۞e^h9 ;%+u<#1\mLC^ mG;.e:.7/yAԲ.)B;ԋd8&SvƜ9[l6'rPi΋(N*z-j%J(>4͛m %4|2 (O{#К58! vMPDءLQ&v(RJ1XÌG"9U$ -9֠N@E)қf R ^e꼥7ޖS|y%3#60>_@[o^6~6`R ޔlZF{F1HA{a^x e4jRBAZ b,%EWVR-o8)#>r^L}Fzf~ꭜ- 47IC{w%=Qo—ئ`Rk4U NMڲPcܧ$)X`1#Q)ȉK9/g^}?d{Էؽ韨g;[1Z40]k!wl/mK_fPTLlu+[͏sB}0-h@]MPH??+U\A߻'??Q\ˌŒJ\R cj]KgрRFj<'[Hܿ"h'<̀j r5 ϞC{>p[?w<b b @YbzN\lRo{hGEÉFk)o c{>m9ZkE;ZKsٳZѢ(,- EW!D5`9ώ_:"2g\.S V֞MIi:EYeUoԾv]t`M)oJWwg7yfz9U37hfU̪S5Ϻjh^]YRX)CU'm_{{Ӷ:z>^hӊ;%+u#<^&m C^ P=#k=˖ | !/8eVzQ(y DŽEmVlM|;Oa/JWZ OeZsf'zP!ƠSV dRmB#U2[n˸-d1p{ C:ⵋ=Yg?fmñYLOƳ7`i80f6alɝA Q #$&r#ւ-gU;0Q KB*;0rt< )D%D^"J@% Iу,|Yu1`[F5I*ʂ␲"QFI{r:@jZմiCjYkJQYxAI\,y4(URr˓6\x 5"ZT({XKqp^i^jU>mxcCiL'JlОSϼT EqBj bGS_w>98i CN)vq *"BQ0rAAMU!|&Dr!gFz\ޠˆRAd񦦸^9|?!*g{uvm}ۧov?_ɕ$9g({zEgHb!O#Ԣ1Fg; M~6GMWA>ϩJ/:rwIXD995c}?R@L|?|T1G;ůn ׏qb_-IwUQ Ɯt+P˕?,7@ۼ|Jd 1?{FlJ_^ܴÀ&I;`2.vkcK$'q߷zXQKԶ-wS",*zK;n+S)D&5ϫdR A"F%)ۨT9yMSǚbLi։K>|?Qm?[mjJub?+WݯՍ‹y`DlX\~}8Y9uR/`M;"`EkM/鲭^­w0i}*=z1ͤG:{)Z+Ak4ֶ2KGi!LH`ңG"|A?{(|Ca8 W/^~~{o~{w1Q_wÊ'qp(07C;]?yߎUC}bThn:|zeMn#ʘ[M J(~_F.L"p4[5"%u~ee)ݞJ.J"SܥT"Tj*|¯N]o+hn{~7 gE22%w"Hz[`(Im輲>pg =0+7((QDX4 / k(b%HMN6u*Pgcb3>u{os|͎rk BS@\$lxr& 2ZMK)- Z䈦΢B7{ Bڲ,O"'7XytX  Q{1ܗDz);#hGϝR[)s(jx9$$ 6^R[rA)ıH1p<xwc^] .z@Sw OMI6(ZfQY9Zi R<2K]fSwE8A7|E#%Zd6)U8XYBvgR RY|RXEdJ!"\B6x`FM7{X%Rʭա]_JPCPQ"c(/&PE\oC'Wl{?o L/;H:vvpڬpc ޡOlPq-b9x=T +|zzԛzQ.(n/Ses8x1ު3fV ~)!ϊ`\0D;ĸ$ Sď5(w4w AHx3CUT7Sf*Lśx3oTBjxx3oT7Sf*8}M,P¸5"ZX.tD%ԖDRdOaLfx4:l׏P0vDH |rVk/瘋Jx->c@2a>]\ u-j/i i 1iUZoNwLmW!fDXD˅1BQTJ/yP[d[H rH# be}0`DDL ` 8:Hw;g4Ck\y5|9$CBXOύ7j}ɖs|iut =\'F+##%| l=9m&ۖ(KaV&טG"hD( F *$;A L g ^<ؖUtz_O;\l)*r]-=Y+1SH_33 a0/߅qEnMiBB?MV,macx'>_;UpU3Qa I_ډl" ܧjl a'CR1 kf7g9 |9l"p4LG ΋O;oL &/jrwLPA7Em0\kR@E1r\h,né ԰h|*>aLQ-Q^+$ZqS2ui`DNl5T1ny 6T "i LZzMIaA)P\`ϐf)NNA.oiG|6+c2->d%2 >끼X u6W g^K10ugŗ*LH0v7@ڴجm@lޮmko^+J'̟5]]N{(..\CۆQ XoS5=ͦL!֛AggKTqK+WT(Hz-qkd͘&h(3 {ZHmH:(0+o*ꎹڢ[ZfaPۍ5ļk"g]8]1k^)pׇב 6ooX a**gh`cO,V3AQ҇U 2X8l(zӀ9rl(ތi'YhNOͯY^߇q$8 YP,",I KErl sMÞk0+ڪ~˰ ym%.`Qd<Xi$X (m[);:sB坽?20"."nеjM]\^^nX`;olc$'sjI 3ZÜEa6KDd F3@ȝ7:-:PP::k K ša03{oi24Ֆ Kh2%C x@ 4yR(e]/{Wȑd џv]VGa 05/5yJD^Qjyf0}_OdV؀eQE2*2Y"*-Cٚ|AY{6:1,`qo"'pˆ̫G/s9: \e %&-(cŨ|2Euڤl(Ax:Ī!Rqxj*y ~winx8&i=*K-c LLjH)B"P ⓊE)u1-&f =jPX7\LI:Ų 'P V)/3*m3xn3h e.PW"=$m..9a|W ieBSXe#r cck@Qlx`aZS!MK283iG5$BkԤ`D*eA"(y 16|2MC{:j4h UuJ }SeY\ l:R>(,Ii\5Ȇ9Y#@"RcsEΡd (lɸ+ 5"(zgBO[TuLr抩ɔ lʄҚF~7 ?c47QXq'E $yV̀j@o\I1#h`\Q ̃V;vbLaxwY3AM|D`P*a*ʦ[C%:ʩll@Ȇ(9nLWn_iJ.L:f )n7(,UjV1aH(+:/(ة(HLD*ХUGE]% AK&5"Ix^,.z־Պ!_U0Vr!᠄bLa`"@Sn iXxJC@ՊJΘ @FaV; ^[-F%̷<kc@JRvp'Q!&x60 U&^켝s6ϥ3lwS٤ ^ϷZ]ǁmZ$,>^x1u@᭫ɠ*MPdnI;!\]h)f d< ;,(Ԩ¢ cʃcF<@ E&rZVȼ"P>X\Lt~XXeFƁSPHmP_ )7Y9tHp# VP''4+JXmA5SRv]MYYXwۮeC>Vyd*,hb⮆,k[0 w@܆=pi/FsZ#&oPw@>P1B(k)de@ %tMކVhGc ;fCd :հ (N6G!aUmQ1H+kFV-5*/8t(Zi298k{ l#`: gh ?AjءvGycP)EF WCQasWs΢b!ˉTIXs?-C:f0M`U#42Ko^5JR\GoĽQE(,W039d@E>&-K X{0sɺ6Z4;[B+N{uct1/ۘkZee-T\tB$0z`2oa6:7 mKQ#bq׺@!58EMyHY5X`4h;fc\ ʇYφWC{o>*c()R(.2PP(yDAJ$5@Y#9{P*mտ5o:k/5s%`6B 4<,,~T˙P#,+QZcP!Vў iDF:|G |sƹ7 ksڌj!$bU?x{踉-cdaVMjSEH"b á %;)@q9x DBF .0] Ay3!zL3`j?®zc(n\,&V]t Njf һGFC>P u4j}\gCJWܞOf7ROzܱg=W>p 0{^k/8] %%NUdwm:|,NX s2;mm4ftRڋM9̈>g'y29N?}|.z*)U <}Я"mc~ng6$Frmk~fB f}҃u|~oo^ߌ'yߴOrAFm-=.ikኻKԚ74}M/7x&9&99ʚHF< {~s'c)˕G8 瘥GiPԒ8TEv=uBq,(t6SFm3lŹTQ͗Ol1?d>ޖZ}ŷygx :Ġ5;Ŵ蜓޵F5PA?hB"|Lḱl(9ԦI%cl1ȨUʢr| 7{*v?ru]!M>tWڿ;4o َurnw9ԽOs|`M^ on r {~g}5 cJCKm|ާɵ9|ύ}/C w_~X-Bd][ن-|إ}g\]=ҥ~M) _swm2Ŧ{7{lvtRJZrbC.N|rږ0Vn٪&(|>@sWkiy/K4{sm!WCPׯ-Wn lI"Lַza>+[οj_o_ ?} G@G˒G^g.RAul:"u ޥ.k糕bN7zI?4:4].Wro֬f;{{],jl>+*x\ p>Mك٧C٧iҾI)Ӽ}.#_]n]}.}>>c1UEI\Pw6L\PqP_=7?|_ ]]x:@D:F!6á2&su2ʤY3Y6iv^X+-& P+VlR*^lGAse y qʛ-G߻oMaz4_|{wr[ۯ#s/TinKlz2Hϡ*I-y"{7B{G~;M+ i)6R)ATR7gp~'/0^wmmHm%[ͲZeJVSd*~U,VUb0''U3W_yYEzP2SyLd^&U,v @E$*'h%kP/%,fPD\2UT^u-tN^&vuJ['^M3o۽:9m5>Ϫc)HK71L܈(}n\]jNh}pf4Ħ`P.+&Z*#OL" o_$ߪH;-GNEAyDN0Am(=cG\%!,1L 1=罥BD:Q)APȝ^iTaY=-F7 Vp:.6cn 8̿ͽ>|GkJ?ޣ g쎮g]nujK:{7[I—8U.]^B[Wغn]5w9tf%3Yw-+lY͸un-ww>rG+-C6?YyGKc7?Y;:.KoOf ]9n?ٶy)mt{۝P+[gm__6*>m\c+0҅ @D2'ډa`*i#̍tvvF8##]('I@Qi퉗ΤhK(H/FL|N:?@΁|ZUȶAm }xJz&L-Sv47VALW_rɕxE `A^OLI+=+QVsO`T]鷘zR篔y IcSӫ/ UV.y,IRLu*yI(VT&<0GD2B{ V8Zzln9Ty=Y\uf "/CY/dKa)1)C\OlQy~898F`*  B\ET٣7f9QzI~ r:,}iIHWДWs}KAbnY~}q=ifzd({g]M'7g)7TU:&r4aL(qc(%jr\OPX |' l pG*wZvr>-IwA2;sͲ$E._X:8P,CL<S K{vV'mas&p~Ar" %t\Y]٭N˝[sw-wWtrd lX=qa·<9lMz 5<2IK- W=^,:6zQ)Mf8p{fBwd) V6LD tҡCl{h_]nxI9P^t'B7 FS0 7gD>8n eZ(juPM09&|>3Um*xxPlF3Qm]rB`uYGZr.Bem9gˇV}GPTJP9`B+%=蔞w :=9Ŋ׹딵3jףvV)r ђ$@PhN(LGIbY$Rl*#PEQJJ)^B*Q>pijbԊNpolg<2*LZmiJ0/FΞ >?ۨ[W~pk%mAfٺy-Q] -_˃n.Rx@ )K_Dd:P'|lUy I 2_u| I.⸦P9͸B*):iCX!5R BT\f vJYTy*.~vu`,`2-Mā @G xZs_'q}$%ze-Q(C)Re1%GȁI5٦t7Ę\(!ܪ,*(%!W MC^ɤL}]Tޅ/٫yurBwi͟i0 n)!%.Wzեr1W.ük7&̟[ WA_0A3gw4^ Nħޚ!}_L\P@vLa=tqK.p&x)*8Ht* D"(r r׭>>Aq8S>-#zW[{V"ukYznuV|" xjrJ]Lm}8QynLOs5ͳ]uhZ?x7am FPd?Lu.vIe%>XAm$ZGrH7mða mfY>Bj+W1~Гti'6j\z|2>Cn{j^ި9y_:eMJ͖N?q/o//)3?χ};Οp1pFŠJ?G~ۇzE=sho>47b /9ق3.9qqkc[n-@vҟq??O~rl0|{rAQo,I/̺:UV*@B b& =Ҏ |¿䓸>/GBqɑ7n$!YE"4$+ q g\0RbJ 6.z,錤S{:9z<  iMHRH4I%EDr+_ Uq͂`j=XGC":gOU'l<=pc{'(/M:/mU]/A$9NR:lyPD訠JJ*O>4f]DSgǝEĘkϜj*x%D P|8H #` @$5CJ|h)I1DҀJ"F.A_1rvyu; з@ o7Z^8 #^H)^`x$RELzHCs>K.H0P7 FJ`UQ {XT<,1il2@ }Ryt3eE|bh&"$wIoe?) mBHL24ѱ@CF3.1.|bBbx5y_uud\,'gg~8ZXm`<9+E˝9k<~~̵Ѯ1$0c%F F$D > ^$7"4h JEe^(@d2 UO &!YE0) WPQ*KJkbqOi+Jy]X3Յ..Z33gυ 9!B'WY\&OܴY;z3p%+XR82v*pp7W8AQSQ?z4oud~KO>ý4}}6d.GM*G9)xWr0愐 fq/"dҊ#kr a)m),.S,m,M•>s.僱TfW/zNAƠGąx4m{hEu Լ';? \Lg,>V50R b:~0mcz)\el/fqOPZNZog)YwzJm25Ap ~cC@u4F?䗫_g1 ໳5ubS_»Ovjs'/%f}{9M}?/<㯢\\i=V5wrs*FVƆT6TUc^Hw|Nu "ur)\ІHo$с>ipT]Lbdg..eI^Xh`v-Z3Zei룞y!q(9(a#bZsXU:V?ā(KJC.VO)K%1XaѲ""hi 478K""8l#JϹIB81 L$ h}b)ÔT a FT&b#}j3z9Pj}urc#U/5yzϱҒJɜJ񶟯d)غ7t"57| pg< G5Ħ6ULYJ{*"=8j{ll +dK F +)IAˀ9=0"ALjEp&eF2 ,Z#ՁHOm1q3vW|k=Kw/{;Ǯ^_ip-ŇL6 Ŭi.,3)0k4U a@MQq)I,cq\&䴳FFC4R9XV[\.\$ZcKnQ'g?ɳ^GiZ?]_ͲQ6lёqsّ 잶)vn#HaZ {"#qzdh9FZ9Bq)1GDg9veP;˱#R'^ 2'H`|Nd !/<׆{ydqU;5X??ka3.}f<[f7t=[\!׊6qm&BνI()sgغUz}8ɆN۬d}~4e-UvC͝ώyͯyxmV5זsgn!=u6,6Mϛn'k맻6rCtqז[g6gRm"y#ޝm(aTA2"=J+*\C2{2ҡF:k-2ҡF$I x D#Nv&E{]D>Q1/ BƢFI@u@m@7h$ָDphiJ#ATH/&f#}m2w,)=r='~ ϳl.^^OVC ^Eˇ`4 .[$1_=_erxTB;9lWB\NS}eEHZWDT.j[9 SֱؗCCу\ݪ$I[" 1$E~$!TbYOJ=fGm ?8[P?-ldwWMۂn|o`犍8d·/Hqfǎy&%zP$NBek͐>Vt$u8jR}r9 bvy<|dYwBΏXRX)6:i.//L+We'/z֒Z6fj[N^PM(h CR=nLrM0{]2PN,S * Ζ# [N''OU6@{irqms4az̾f̒q`L9s,UZX%IǾve_y~٦:Mrb,jҝ8Oy m}z?c:4Qzs;#2HO,hKQ LXH\yY8ހlZqzsT$AR;y*}Oc_ţ ҦcUq?rhϮ]dБ" 3\-K9v+z4hy8)-:.q\ bI9l&8 G(57J6iA4oEbxԲ] =8?* =ˡԄ}!Wq8 #ϬJ$$&aG&3>xMJs-hQ{{% JgOS`Xn{M愥K,v)ocp/_)jd4s˦|T5ו y eCGZXZE OL3Ub9̩;Cu$,]9N7uiG:S%(at\`R$3H+#tJȉGhOKb$+g+Hg:ot`Ll]ߛ3u\!f^-]oKٰ-r-V1(+NfjgDmhPTRUj-((LLmq&:cZyF[xvێm=z\M^S鷗kcrLTH18!qϔ"ZnURy5Ѡ-bZYDFtTR$%i  2O}z\k҉OK/]!8U4^k\B8;qkWykp:+< %8׷>p.7Fv=վi^qSS7Y"0%%|9b"MHLF'p&%+j(2JR9!qzϥUHƱ`c)U1H*TMR8q=K9 b!/½pMobg9[n⧃oTv0}_MÅu-)' G9xNX>doeh!60s)ba#C$LPH8&10L>&`QH)q6# Âq.6;vEm(Сv`0@N0˵EjƤWk5 aɘP1&xT;-(̐ZўGX"EDXģQ86r[Uj~슈0";D\"AdBLZyo;ɷW3b^a0\aizxvDFqaSd*Utƻ◇1z?!Ћʬ 5s]vjFsȦԠ "ti3+^Q>fsd}aZ>V|a')(8xl?u>7[  N[lBpABB\ |빓"8;t$( &|NJ >3EsMaD'a;o LrT2S75Řki7 F&?f4}:+O!Fobr:p\f a^ř~i[x[ eŤ8 W1G hM-q[:kj47pc3uayJ`X'`bŇ6{gmSUnouɦVW,0ӑ0pҧ/!,a?n|`~GćcAy;sS:~U?~~ūw'{fR&j* G\Mt}5÷fެiXSiWvvoWlUSn@~\|{1/{cv]*+?ijW2+@6Y)՞Q󂪿S"U&Di!f\ّf|P,@E~SLUZ~ŭu4qly ' 1@N[H@p]L!-QBQy'bDHW vNCGf6//ߛqxpꥦ( oPP QDX4 k(#FKǑ8 -tQ?:wCbnHĶtt+bj>Q72MW72u:mhZvgSٿtgɧibF&x;}"Խ:{rC 2q8'S s&$8E xd'(v݄ )O=|P*B {ϣ29{ W^7l8\`N×R݋QX\Ar=j%_jgg֣R]Q9Ћ2(PZlp` =+&kTߣ?haI{> \HvSFHC^5?[NL|?]*i SkwկmyxXʏNV50ݬ U덝4v~~LrJ)`p{s6;˜X.Y^]uui-4VnU iHd{BLE"91]M#LK<%`E;4Zn˶VӍPmOq[T!E*֘D.Wbr7TꝹzJcʮjg )(綝NE9&>̎kUـ!~! ?FInap;"#C@L!\ VJ^Wy~HellUg6XeB4pm:CP-6'c054 3ϳ9o8m&_ӤSmNUfͱJwv9[GL).)$%qe&`^} ݇7W Iwۅ0n .dxJʭdUb=D'΢muelk%jtsTLߏR-^#nD0Gqݨ'ca̕ڙUck[UpMBSy!_^ϪnkZY.prʙV&W\[,Ê`Zգtr F)!xQxT(xߊ?HL{7GhrPywpx^1Ϧ g}R1L?A\?{`c1R* /{0^kt Y"D92k:-+ɧspa3I9J& $+芅Q$(XrhJpvqbvQ-HaČ!HZ3b!%,d!Xk,}5iv1(T.ꭏfߝaEV@-奋P 8;ĬW逰NY@SJx->cmLBߔEUBR4PKGv g?L>|4rW> »țσACDǷ.ZÍٓ+U'mz~|0.ؕ poփBYԹkYZw1ӏi埯2WU'_[Cm4ͩSi)ZdBzXQքoKH/QKզLBzO1K19se.Cix`,KT 3rxssHkwW*JCkU]g9LU jdJGWD~E:s82D:N=QJ;_yo^#ؼAڠB_•ݜ:@c7NDs!ugp{#,v-u*{qΙ{g` [)j~?! A-kCV)^wﻈ|؃o{tdEFl_^|BYwQ!ݭOsӇQ 0EN93X?p2DCECGE4X1"a8Y V>9&r6ȹMEU#Z)"VrH:uF߼ڤY-إ[;A2'GR`}z>RPo$nܘ{sf)y{{Ά|2d R#1'+GH.dahpB GK[}hcrl{,ZG+N !t>`gI\*%X F`<-g^2-B$XEʖG"`4)#"b1h#1bg,Ge`k\*r"^} fQ˔.:|uCWmEY0>\y  =_DGOV$#%0#Wz :m$Y\C(e)LժS vh5Z`(3*h2:ڀmb0XJBԀRE7oNY qw)-ylquok=c9X},!;qTԵŪ[ޙ՗- -dBk&knFo[a$OWC7UMe]. H)JPN_CR$ISD[rdY8=#x,_Yjn,EeU9x% m҂cM`:XYV#3GNGQBN Q2ce(L)%SKS,.4)2,w%902q3CF BNmЕqg{`m4 \$a<ׅ]nHeŬﯽmgK,yqnzq լVwIvƩS5W]^?A\0^nuu3o{Ziӻ6d6f!leحwiy|x{ z^iy?f;ilsZtGo]Ŷ|evMaˈ6/c}2p*ܝosQ6pW憛d൸yWy]rgAҥandv$9JhO ~ .#H< 9YXɤ ,(39u2'~Ɉ U\VTCnmUt4MCBJ/ȬkK?W!s|L#賋%[ywf+^QpH_$L3Lրb3L}`a' fW]N9H%uw9Ejvs%WE6ZY0@qO:$<8- dB8[! d1N@(^$X**;E,jmJƙ?l/D\>|-iJ +ft5sJ~2[ǞrO&Z/\Kr]* rpUĩ4LH?}4˳r^<"/-?.h0e)6֡HG`!vץbRv.&Zn88|.Q,m4asǖݝk~3+/uWYe rsrÍ̠#4)3X I+wh EB+/HJ7JaOK!7䱀OsK8,}yV[['Zb|?b /5>Dg\\v{59 ' =ruʂ5 Žah'{Գw=j;(6:Rkq>etqEvAQ;&->h"8\`<o h{9A&zŽ%a~ؘLhSZX,NY vzq݄wchdۡc&e_wBúL:׾kr34N7<$  (j o%I{=w8>cO@b i.ĘZIH-!D]9`畧cBꙁ;jo])u=q &7T6:HB zz=DNd)D/O<>!Xm$C8(ZI!e!9?nGt~tIsJ-?]/Yy@>zG`eFzQUH&DA`I,p.YT%䚓%P7~AY¶2xmPIc:3:R!x|Ϗ(xad:Z0R:n9[!?jy륁cyУ]j\ri hAI!+Uy4Yv3[d Mܴ#,q 4Fa4(Q32l,^f뤊 ID$I*PmcJ4MoWXw鹋+ %36J!FLF`fAiA27lW4]f-s$CH_s. (D2L;Cc"rPqlm3R]AIS*{)´T:Hp:D,^QZ #4*OxrZoJlUSj^i6;{2ICxSZ^ EƮ]IpZy7F {Vi5]5 81e](܌'8Fd}%_<ᖈ@1LQ:b'gFp'mEܞ3R'iųV%]>2O$A z.Һ&;=XvP9<S]Nt--лwiY?jH `໓S+rW'8ծ7tN?ènBh=7/&FggO݃nf_fk o x'ڂ_#hz9({~w;IӿFIޑZ?E0{Y,b^vVb9/Bc/ƳM/gm拓mu}/i[.RF/cM<*j㔟N*Ǯ>&knl!:!~썓$&90ĔWeR1ꜳ&$@Kwl I/mmX꼦/hq8r*HKk²FϒMuLfA;5Rxgx ȳ<.u\?U=?gusv>FK5j_z%(a,͢ӞWhR<598hdN \&b$;~{izǴDž}] E ~Ӄپ X /Y&@l4֟':L%R2,w)!)<}K љ sw ޸t7ym$/6Et$zp0Y7Daw GM"$3d \PdG![/a"d 1X[ +[Mg.#8t]*X }ʋ#I(?]v z\慝V?͵hOκLaY.B\III֌զCDF*WBe].XaI !Ģ#$CZ$a&JVR< Y$Ralׇ_x4Ej}5ՠ1HiL˔pvICsi,Iт|,fo *J-,y'% 4RgheIsHLZHp?\eXm:;9'vY lvɾzQW֋zЋ^T IȜ6Q},/!Y=CcчvǾTևf?}xVz"EqjoYȝ5Q'Ht^;x i(Wio翋 N9b75>l4"+փ[}dG((]y 0M&(9eƭg2Br #̩nt4(O]z`wm4eq/mf ~ֵ,yp,bwGˊԒ嘁8Y&:RȰ, Ƹ#aREb$ UȔ}y.hC:'Ef[6q[eg憍_N'oA08G1R02RgP™07M^uQN 9Wޙij8N`w.!Lˏ)Tu: ۉol'vva )5BhX$dJg՚yG9 Qָ|\HS? m|ek[niv-’WF-P_u͝-?k`fwԇLV@oE/dZiy.sqNGsQS^"lşgٱSxx3(gZRlc @GJe1/TLNyͶ+u.p6yB_9}A/ riHUz㑧P0`zL)BRCT[_aR/1",DĊ1тFQ4r2&"ݱ/3q<8ƕ~Cr0'rU\_ ģ5Ҋ_-}){}Fˀ`M`D']mjP06*X(IF؉FHcY1d x9ڨH'X@)0ӭ領:K,(RFBaO{IUC4| M 1 ywTMP]K.;jܬ-ݲi`?ib_ <蜑VD$mEX=L((u Zġ}H<+np4Yy~4cĀ ( [S wZGFǜHØED$n\?{9H@qWtЛwjEnja0^Xf\$Z!Q`.~v(6?Q2{n8*h X`K5OVWez1Eča 튆fiN`p $!__n/x͓ ς.;V< $0`k,RJF#Ř. R*9 R^1^?޾fGν6(^9<[cAeJq6q":W-D!c-J'DJF8LGi~| ŅNJ=?e+ܾWܾ7sfn훹}Dgn훹}3oܾ7sfn.2oܾ7sfn훹}3ofn훹}F1sfnb֋7sfn훹}3 +y8lF'w[\\FAOLU2YE0`3X{.c1>i -.0~/?ww33Z7$ E WKQF EyX`P"#vKIf\.% A%tViǬQFYJyCp-a^j1L5#12s`q]I+aiXj#ZgƵM$Ҷ;^2-Nf uj ^Hi+0,aHi@qo׵6pe@`Bp{ 44Bq PLMHRHNcRɜRafXTqXŒ&B!BՀ{<\?Y앦`wÎ:Ӕ9 GdXk, y?\Z֙H* еZ#sgf#OHHh 3G:Dy9fZY6y"u;6rP !-N`x} ˨QiPxIQr(9T叻P k;Qjϒwi uZ5Lt&1zrOT`^8M'T\7yƟJ ABl."gH&8qcH)abxlx<xulU`TE-6`MHE8ABB*,0$1}zK; ŵHUzmofkꫴdR A"F%حU Ա<Qc E-`Tz635xr1+WwՍe`0 Ű2_{_}KoD`8x?Ty wn? @kk֖fH{3 6a0`ŴYGmzUTknum=+X.tvl9 G.U=b6N1.Ŀ|P)|E֕/`8 ٯoyk?ǻt:{?o>V?ALKADX!v-x_߾_%ŷeެiM+S7NGni|ne|ɭ&ӯ?/Kj'm8-VM{ p%u=}Ći~UYi5UTRTZ 2]J(9mČzTP@o vg7}>&4mf]ʍ$89m} qvQ3ʈ ;c$B@Lh}{f:ϷC8}O< "E }Q "E*ba EP1bzڠɦAlLlQZxÅm(2{K$$!d7K ?KTxr& 2ZMK)- .$9f=Hlc*"yBHqgN%rEιW}IRm1v)>'CJh!e(B;N0 71L5Ғ4s'_S |hbpNtW{kADdS+S4 @GvBXcI6z{AivY4*o_DP{&I;9.`4h.j2$zX|-^$鉝c|Ϝ`ES+ä5"0Ze[RJmɌ6> aLf_ť~w Ϳ ;|5*C,%@: lp6HF7g)Oc.*5{bUVaPwk?Q/-w,Çm6 0LqET;S" L\Z$}iC6FbGNRMQ*tɿݮ{GG0@OH*f1QkVDB=Ti-7HY"?̓!~Z&LYfYׯ !W6PŖ&UZ/Ā.&hIJPzTKs0~y'rެ*%1_}Pӕ7n5j^*z<[\<:c~- ;ӳ Wuf{9{Yu#A[6M'[fl[׭iJVzܲm\|1 dX'Vsր8x=WμRW[cq4*2iӰ<٢ӧFPmJͥĉsB:AΔF_n|fD>,jvfBW%̄D%ͮ؁ANuP$"fAHi>5cGlA5((.B +7‭iUHF[ :aFhNPH]u%ӮF/nFguayZRdюjfJ$|o0}Lwj0}_ &gBՔv[4;%{rBɕ2:jWr~>]IcF1K)`)$$|yz%%0P:oQ!bo#Ա#đ!sSgg vO! Fr^V>љsL-0ܸW/WqDn9uqcgn2Wp.@+'~ I{A2@JF/_tJO03H1;=g'a֯"z+6WW=&KU,R=L {Sz۸4}0*Y/:P sZD)Z2li%%iC 2}y辽Lto^&) oh޵#"`,/}Jt]skttd H醠546phH~3F3V>LG J IiV˜,o"xH-B{ɍTY,\n+I3pI ldmY2Îbg}6.4!mJqqۜn'oqjnظ٬R$sB۩,6c{I9rюC%`V։(])*!g RysƏuB%!MO3c8WA GLmqK]WW%-˭`mk-OuY0zj/YRPBK"XIhY4]\*9Ո+cċ Zfqu/r {"4ְW~+*]*Bx0E`" \ﻸ*T.zC⊣;T^=htÏ9oJ88n-̤*{D vR4)` l5֫Р%XiWls-G1%'?'3xmIxD?z!IK*CQ|-vK UIˣm@K3J)rϕtwJ#%]@YѳexṲgJe \$ÃBAYSJ:f1*g=$T6&2PѡI}`V;4*RԷCwFQ3O x>z~K.? Vc3/ S|3L `fqf «R''  h>wૠ\#ce- ϊP:N=OjnD6dB(+uX.} rcP@(Js5: J\ `i4}qorrZq3x^Ҷ7G/150:oå cw6!8Ł5փj$SV䦔&8tM@VE9SGS&*{Hk=2.cΠMGHYZQ#AETb{( yYEͶ/6qt}zrʦl4 J~5锒LNj8K@_䳕HPl!kr;BeO"^o*`x#fB!0> !j!xV CRASFfC$2?C#`4FpE)Vy0:]DBvG Dw{7WmBH>;:-ݭbz| ~[37v5Y$c R /Lp$LX9LFp5!*x<,VѳVT$#1CRY;$VbZ0}Vtl8)=VӉ)p݅ARW8~aq\O5X#wQ{+ c16eHZNekB3UGy r}g\mXLYoB5<+80{5N"qw";K]FwPW.ƥL O@eC!'!: ̞8*fIxI$ T1ޕuGz2/O,>oxxY8wSJ\ Ӵ#і(훭;NJ3]le AOL?O?;OŒd<m&izqH鷶骊ϲ7nenuɾE\la½UꖢTx2*y<ׂ# _Ys(lR7m$Ac"'mA+ɥ\N{#"}nﮛ2`c}}xkM,)1qXkaI, G,5U)$ZtŴBYB\J9#hh:PI:}49" ͎zgq>vP m8x[Clj[mza~XZn.rK%YpV(nd&H3}Io,imp HB/u I|K.*@]:n4QaA~fd0XNfRG$GPJQݗ Ua[Og֝ޢE!`@L\AB!P̼E)eZl0\G.dF&3RSr.y<Ӷ-#=J;TRʆiu B,$WQD$uTW6rv6r^?+&*—b({urBI4k47(7 'ZJqL 59^kR!,8)ʎ0'_;sHoKYp* ά@O[d}MJ 5yb>ttΎND8rIG) Gi_YrvUƒw+AcrWKZ~ZlJi!@ZG.sBV/wZ!Ҷh__ku2G aRaJQة  YUvjgt|kUƥԻ1yެy11ү'?]L_f+4-1_aOfk+zmDhEwn$׎$ꑾa7GbژV> U,h8\,zt];*gG]Lrݨ*5%k28/&7ʽ"6zٽĿ)SM{^蟟v3Ӈ_ǟщgZi t(#aCJGyw/$94ZXVt92rkƽ>F]*pu $ħ/P<$.wOp>5giS 3_!Ϳʒ'0oB=RVX\Lz|^@eMC}Mb~tMֿq%IE@}L<1唄QZVlu m 13'-oYUڰyM>^WGm(ut,YYgdvNh;Yܡ#d":V/u:+oum<sKK&]V2KV+);<^yGM*&X=FޗlIZR8VNB6أb!b!Γ'7Z0 0uZ:T&,z1gQk!n&1`a8Q.jGàH1%k@陕ZFQ.Eh{'o(@ ` 0z/3/o\"1UYϫԝ]OgŅg%H+Pg2gG4L-:<ɒ6@MUV1sX&幷ܔ*#Sd[CnF*)Qu`G{N*WO#'JY A^)31%c)oH4{KYms=q݁UKd[AB@>qK9^im7ExcVR$e629x]SB _Bk·T{!=u_L:+$u`.k!3ɘʜO!i ؒ+*fA= 8YgzH_!]vc< Lc㠑1MIʲŢ.HZJRT-Ѭd1*/"xbukp)_fCXOwD(!"^1,ekQDCtڨ Jz-@aV{{ s̑jZ{߹WjYd(W~.B7{;>ҁGE_ )}Ϸ:Y7w)x%V[>laxmvXLTku%,72Q@|ScyKV޴??Y!F/2ӭ[>ߞ&{͏O-.i8׿=MU]InT9͋V2$IVV@γv#5-y~q#zeEF O|!-:ѱfh7.ZnmeM"OF \oOh#Z>O+}O_0ۯrUZi F4uwoGog<.%:o^Lu[j|\)8Vgh',Kƥlp7Uh-% qfH{t2 \ F挐$MQY&J`O*Cy4rꆆIɧDDhrR$Gt qYdNv1)QMü6B1uDuhqQ|B\ #:Aӓ1-d0mٻ2W|0Q-.0JM8f@vDZ*&/ & ዒ֖_~ 'j5qOt6o1oc0/Kv5=yl q$.PJOH窯ǷU#5p<=N}Vjlx|?#GS)(՜Y b2x,qRBfc Yx,+mM1#AYrAГ VS9o+@62Vgj\Vӌbr½n,̸' guݵ؞sɢ4~2]MƋ Gl7T:%sd6r,Z$ه1V j-AD摩ശU0݆b(DRؔĐ4h#l;Rf8gx<ԮEm0`7(Pb/m d9q2^@-HS|r1Tì*r +:dkbbA$lDSL/fY :+a5qakԯCd 0M ZXDTQ 8 $)cR&(9d` "hAkFKL0hbQ)_n>I.ke4 r,$472Ȓ`7Bǭ!<ѣFEMcuVӒ⢮z7q4KrRȀ9udX)-*(>HR˦յ~f&`{%397FF/lnNtQ w#N|myJ0Mky:5ҝ LjB;LVaJe퍻r9>Zi=৥U-ݨ>K+ȓW4c6-㰸.sjM'8#?qb4l ),2'2BW6+eј_FO _Wa}N+^n؀wq>9`PZYj5PJr~1OX̠{ؾv.bE Hg4ZUrz[*#>E{].nŪ`jښy:/o/B%>F dz+mJ@rKpUU]#GQu)5"" <8Hdb6(x$ T#zȥʴ#V\\EnKAo'iCHTSe[Ka.SiGf/?_#Go]ۇ7퍳)* /"^Kd\/dq:垎9VlK`Vo!jyf1uX7:}wFY`qLͳeb!X㼵 S &Ԡ'#WX;{v}&k2L_2!N#IS1ż(ɈQL5׭{i-FK{%&^t.jf$)&&b{ف]umR7 &hFx^ vq`GĎv{cGp?$Hhҥ4m*t|USh766ڭ=={ͼjCWkukC6.zJEop_,-? LFOӍV66Q4,Oԁ!@=$e2Ƀy\ panXRzA58*PB3Su+*KԸYV u2pC  Y\! DCц\|5q{1yL›ir{q n+53>cEp>D}@X}8UYE0{G>ϦrŲ 2ƱHtCԚ76h?9-3 46cG2&(dLEȘt3h]1")EeՐ%לH1`u2j T%'ь tiA$oe62 8;6ꛣ|1DmHZ]_:!~b4CwwCl|5[M_g*vp9 ,#Cz%io@ @ !gͅdDaQR]Ip\eNs>I69.cm<7F4r`Wg ~gy<͖'ίɶ<}ئX@+,K(=HOBؚPuT(BR n^c4#-*xpNҢxĆ84Ü258^wto~x1 œi440I"hǽC4<Q|y酞y^xk3d)WsJp8}"8&6Zo<2tTAB}?Wq~wό'?IYK[If< l); avd|QJT)ƏaW7aLֻԢg?jB Se7w_ew~<R^8H%Qc<'#1: .# ^ QG&H3|Z0~of1\_~*+v7r:^.,ہD[^+]+R]EnEֹS=7hdOFcn;c[ԤK+`>ON]2x!j_>_vwnٻ .%'_ xI5q> Z9oYk,;$ bdc,s Pz6Z8*!2*2FK 4R"ל` )KݲdTPV(c8\Fhh$xAzg*.Z}4sD͖Y1Zs;)=  %PڴeFgWzwH-VSKr-W1u 10tA;%3 LE_NYIΤ NпJp4x3DQ4VL6K'EZù!H!iPENzNw_yl4 EK{xa]_q 5 lE$BL y#GEA|9iAY^#2M23D%" hHrL`@\TQqtԮ `%E1Iԁ< ^JEWDH*Q97-'zV.VagE]x_W\H?W?/c-%}GoTH.ƷUfW$EGeϯwnMOԟU,: YpE'^ `muÐ,1-< ☢1Nvxj4ptszzn,9I\zcɻTRIɠi 9NpP9os+G3O׀HsEڞޯAGj$0aX*rv/"p.F Z'ޕȚɕUMHLӫH6'g^5']xB6Uo9s{-.8pEsb)1ڲ^:Ttl|˗,Ihc |,XV9՘]gQ{C.uٳJՔCHXs8 T sU`j8٠s/n*㦚xz*>#v/^,7N|}…=yO^zF+?hSEp . oFـ//|M͆+WBO=U5m>qm^/~l_ơD.^YtjE+ WA+D6Yc)ݞ0EC?oR9U,6b۲3rـʤ|/N}wG׵ͬfHbR">&rJQZVڋ,A6tYhDKu{3ayu#>a3MLut,Ybf2;1r0,H#"DtieSgj͵kFrqn$g Ywkg;Kn4e,8Z"Vsfo h;!X=m-ο[vgP|(wdV݁f'e2O&>4B~'B4aķW.yz+yƯ/2MEΆˍm_/z! ^lFۥv+_9ECf@Q*Ax#g͂xoēxdXM(Dk&K;_;ٝ7<ҹ>W2`urf'xz&o|6EVrh01 P=jvbReΫ Vč%sR?{%r9JH-Z߽u 9hq_l$$k˴erP{.~Kbׂݙ [0km3F;?Y}V]T٦쐡 d2GH4^߼KՕ~lPާ@էwiPueDkBLMaU]6Mr}ݛ{1^1x?~,8/ߖrk:zN'fW?S lrQl4aQYηSt7g- |EÛNۢn ߩzi9nMՋ=w:)ܑ= %@8M爢|\`f5].1)M}WuFkukĽs`-R/ab >)f C:$SQw0 Ӻk^/NzM3G1N`m&Ʃk8jQzQ Lw1N0IH+4#*# \&$P+v c+P\(GpU o튨WJ"#\I )lU!پQ+a^ i9\=RJq ?+"W0/pUvP)lW2|B0RА}B-7WJ#\vE ! 7pU9\*ug >JB qz͔iBF12{a2L+8fP77ϟTH٧3_%+L՗~l9}Bi".J.Ԫ ;~($U!x fo\vZpU\;jኯ!򇇫l`CNP.Qkfp;Zs0J43M4cF–t!5{zՖ~18dVɶNxpʅj빪%Sr&8ƴp{i62Wi3ȇ"?NT/x!i=*OpP-Tzz %0 e!J8{̞mZoL:'0CЈpibSpSn[Y}9?0eїzU> =^mޞ`EoeO{0Ih}7\*޵#"6ߏ!白 ng0bEYr$9~nIl$K-'8&lb=j}ʬY'|y{nM~4M.5XT1~!>/qjDćA2gU8l9:ӊsÑodGзQYB;e%Am\ŜxvNWy`ԇ}"ĞRvޕH]~mofM뫹Ǔ˫KM\O㣰U)lrV#)et޵|}¨F;?4-k@gzr>pV#( /ΛvlDpkuo>}֒T- ˶fX{368Y㢍0 G0bŇlǷm^\^LuQl:j[_vH9lHnXprc S/czD{& yRFJΉ4] ۽pqu/ȎN>ӷ??̜woq BF ٕ?#@A^`|{6͚FlѴ%{=m%mӃ[/ ?~v7bOW9a-Lj5Q"Ȁ+A#H6.˹}pVQSTYE؜pkb؇|3v|Pj'ᱮM's>n |V>ZF H6kw!$!YN 8J3.J)1=h;K:!驵 3W!j 3Jn]uX Σ$>HqeR'X,欦#M,DHwSPg =ok@އ*͉%EZ@6 >hwqchH~y~e\Յ^r>0엔!R L!/ƨ&׾jzFՙJ:ѨZiV-\M+Hvܫ$I)k#O[OCL \.*lDfHuQ9b㔤`рJH%DF ˨,δAWE9P&0 \{=MF#0⅔Bi Q*bd\hzsއ R~tmrA2"c y#tA =,Z8:,I+]ͺD Nj5HrV{ c^""1\ K2.,J*w^WjvFSe !`P0^Jd 2#$8܅"D(fo7sL4G{/bښZVutzaRS/Wϣo| o$+C8^y\=5~+cz|ArRݳ26a\@r؜O'FxyQ MQ,ج= *_T)@H\D"9{NGmԚt.$g(1S)tƒVX)P yhxX>; >-x}!og1R̹9Z]—۵R.O}&K,CAsiU=skM 8.D ܓqIߢ"nYS=]G~cob|\ yCna2l\ ;q,~of`/ q*0뮱A /.-0jRXTFz!RVg:+a@hTJb+/Id{Hd"`MmF!Z'.ў vl(VE_2g D@ -eNj ATp/N [j-Jrk)"ʇ/SWMh?a!:&~#ys&X._pP[zO ׼m{ͻ`ebi_:ƷG.&K`['xcVr'v}@gtk(c2j:>TEöB;L<--.xU@ ׯb:ƕ^e$%d*aނ vc[assplm!!4q8b948Ө bI9A'ϤEU oboEyo@Y"1Z(Mw;+tTaUf1r4X˲Iivw,0P³)a"*V` '!&FFJf`Q<% 6T HA@2KEp,"A"1=Y|RbBZ β'c5BX ʣZ[_]xy"*))w5"LF7uDc.dw^_t;IUNm) ljܿ8s-,Z!022ӆ0O,H bPTkm4UhY et1yO[ Y1YϢhRPQ*KȤJsblT fƮ慼ㅏ M32n9onjq\?ҿOsMDKk&/-7\NXr!ۊdE<*erll0 ,q&GW"jfQ"FeĄaU 299₈y(\ؕk\[t\Cf`(PTs0!$̈́NS[aX-,_G8tfɮ|Q拪_\yBsb.1T i$XT$V΍^}$ˇ<҈Q_| x(ؕP3k|3eGn8D1 0J.գ9 w6!y&N3T_e%brDvޮ8ΪʺR LY@~W-iԋp)دh9P"rZ[bPGomTDɨh:wn.e\ZwRO|N6%m i1FZxw/T: gTmi'A˘J"+U¤04C[]p $͆I["  $}ISbTA$u%8Q5Oem <39tww(I YgUa:+3MB5{J5(~,紻ͧ",j;tL[3衦jTB !&~ dE}Jp+1 e$}*;ɋ_@G:Stڨt+e´L-[i RI"(.&y+qQ%Y" ''U3W_yYE~zPhSiSkivT1 ѹ @%& qBCXz)g9="ǧzQIIni}nmT%/B1&$L+*Ef-].tHZ&Kk;^qf^M6TQ[q_(|KK7<Q/$U/ͅsoНpPMM -+tBhM+D:w#܍Pp7vT49RZ)e@g sp6FNQ"XQ R!2$SW],:ךS|"?,Uc Kϥq8 ˬW/Ѭ+Ēھu{›+ +[ԼPr|Ⱥfwo͜*V⩪Ny45gݟݬ_4sR->;UD%NLDce"x9@{s\o/c !]ILHV@;PMEp);!tqB:;8!].BHHbѓ@˘PΤ(K(4 H̖:#U f69f#59 3`5`(bNYDuEؕ+ˇ$fry5~ Y~o1ˊM~V>^~JzL-Sz `?Eg«SE? HEv7 ?{WF_viG2YOf X$0xȒW-'>ah-b)npbu,U?VuٿXLa Œ`.IU]Or#I%Fù% \`0D ,KкVBNgꁓֲ_[ N MS1G(־94Z㟕UkN??lRo=)"X5O rMN| rD-'8M-%w}: ㄲDgQ@csTSvA0I89!oNE%n#*ڔ \Y{ _|TM_m7 -Nz*Ҝ_16%N"E]`Pq4MR;ߎq!V!;`JUꄚ@jB7$ 9gIШP;W hjĥ*c@3CMuE 'gRXrBs b.Ŝ,P>kW.ox+{7%T&8ǎOssVaq=Eكs}S^~E {LOn*'yB!Pk1w4D<,tS =|kc2"̠$DVPK(T֛HQGG%^mpQM< OV׭f?GckpPgt+B%ZN:&2.rd6i璈\~05pFSTLFg"$G)kY`ʠ)`-uG2/yJ|Y=n$@{MWiOB?Z'G*QO/37%ę2{W*{PrޏnxِC!jl$ewfGa8s|轍*M|UjYk=VLۼĬݼj]s~QLAPrzcRe(8a]3g̔{T%i/EZZN`׋d*iW%i2Z \erR+(uT2 +]I+7pj_ p\DxWH*ȾUvUXW/@I /pf*S;E•~~`5m'ώ7\TWQӣBu0\^0q$%x:{5lqqCL:HC#rgk[s5]Njb^iG,p 2ggqa>ƣx'o׼kQ1B <S`btrg,S/0`VUk0%J$x(az¸$p}K985W^|{A?wQkQrSë"hai)<1 <0h)VUn4˕N2sؐcS @mu+ݕ"b@ F (̊hX棅d,L]er %m[rJ*:KZrF+аGp ?p^UV^Rb m?`O#iԊg{\'jݥV*+EUjһ2%>>e^]ZbN4)JտO kԊ%V0]z:fn~}￾EhRu #࠷I[ G & j aD'[i "*{-$@LL{Z9 loFb9 3ݤ^[Qy 0? p?N-~Lg3+Lq*@*gYR}6d zTo<ͩ~sjb`7Mw0&vߗޭrѳE`ZKJ4#&roQ;OJWZ OY(J!$H>Z)C^{-THƱ`c)U1H*TMRm#ck<*aak.򖱐wXzˬפ,=5-J??{pPNq&Åe&' G9xNX>gʊ :"Ralwqa$IѶC[810L>&`2bFƣbb j6Сv>%p@Y\ P4fLjT KKh@ IqǺxZ5+ o)p:+Z푏50QTN :g~4THGQௐE(LJ"v==Sr9Q%Tڋ`t=עod@UPۍfےG}Xf)UŰj$ń`8a9$WLEA"(GIW]LrhHL(E쥌DEiYm\D"Fq24iEk4t._9S~5% S2n8Ԩ56Ǐ]dΕ'5nb405rE{&'21.3c<1K+e\8 2)~eVh% #2* vAᇗevo.q'x97T>ȼZ^X;x1̗1GqhtzTd{wUӢ~׷3; co~1~ f?.0/DC!|T!ѱ=B: JЖeI8˹ulC 0?UV RQ-۫_*::O/(W\7z \whpJ]muYv\Q}CƘof:cUbk8_4斛]9$ƣ.UVcd2Z]OBahZ*E q4 g0bǣb>CӺWFu1Ⱥ^ )-ui\HX 8Ty.^>=>we5J?g F=?8Cv ɛ~~}}_|ߎOw LiHXuIc3ߚw|v-tyOx~UCG~b[n%@v(0(}vZT~zDx ף (z~DEՏ2ez7_{R}3Il`yȒ'3&)Yu&5dXlvWUJ6߄҄h.z|P*˦ńKZ/ImY/6Gh(~ +wrƅ"Vɀ&: 8t˔E4oms֐I:tafr|:ňÉp 'IKAd\k*y`(]`,i_4Rg'$66&xJI<豯vN>~6MM15 N5 +vjT9Ǥs)c@넇h:ɰlWIgDqCjyuy<}5wI$4Q;! H(3[BU(OВivm3WRvX0hk\ jÁG 2FC9_y}51 B 3[D ?E#B[ BLQq\I'YuVj.mId 1$!G*.@Qse[C)&G5 >P1@#0CM YB }1fqT9*r;zpP1u+im{tMv5lŌrͬ8Wװu=h|sh›1{Iu+@p\YˆĤc$g9XTsGJ&9~  @;y5z{ sݑnJd b-Mk bZds .\o[H 7JͿ ]{.ܻtf\I{0L$dXp]ؐldm ֹyQӗ[Ͼ~zGZS ]DB-ME ?ygLDC5Ou(uϭf؝6[nsoe-guzv{Hk]x⽃䅖h0 ovvs=2w*4=k[o*~ܻ )O9Y]Sּb!_xm}+G:ߦq߼E2x^>oO 'W$G,4 l$㭓''NIg' 9pN$R)G LΣb^(IV uMjG#u oSl.-:::!d-tC1 {M$@+"+P%Z3O-Hz W<|׉l_E_Agf)m!Ց ue8M]<}>>IaEa[*(>Mݪ"ҁwN8.lx=t,x# 5/W8XAv CJY*N:4U ??qXh\-<isp_`ՙψ]}}hlRo{RhGEDBk )o -kGA??wڿ/yp2KqBd(\199Cyz-@9'!fQ})3\Jٙf[m8]%5깺:ܫ*ҜdDK^\]ϒHq.i0u$M.Δ<\a/!z5 G0 ,P*2V'D!B7$ 9gIШhˬc`ڲ~^I^,~|=n~cDzFUZZi;v^e)Tl O)$r2nXd'\N(IFC88c$hS%Zӌ&ס5=kMOky^Zf̠$DVPKT"֛H#QɃA[earZF,O򸿺G#? -dz8X u6H.be8KMZ9ǹ$"9Im;wvүb͜42ڦ=ۮJ60y׍unƹ1zq #FˆN*3[iO1/!4j땘ud';$iϭጦ6^n# fє,0euVw)DgxVs Wչxy~[z{|&*Yg88''ǜB!QO)ׅ CIO#)Sk*Ʌc여9sKJ>nae( nmoxIyxxJN-.w;] ܓw.%]%ϻο7cX'}DEvycE'\7,։5Hdb?G=ٰ@D|.ۜ5r5W@~%l J~zyz!cj)vzzӛ7)|Q#cJsB3IA յ.~~~R}wX #SM~W1N.~!#lYԍ0sᒢ1 qԅ_wd_~~:m~vsGgJno9v://Et.n|us0\9x(C c,:b<*ܟt{o\&9 ֜ᄍ(sw&=ӊAE'yunpgrtb?F`sbbGW~*KW,wu up0EˤoۛcƿyhqS< ޫ,')fr=OsD/`R8cy)e<0wo(/'rʫ*.3-`vesJk v7Wj!ZOK2M͗2,_N&f8QbHɢ:nu(.Uٌ֮^rFJ @tg 1B?곳8kМ_NUV,*dZ ";abU"y(dgN5P.Ryuå|$+T7ܳJuY5SI[a%f5I]>G`{eg"x4_g%랛y{uyمͲ>JЧ%A c#ZSbS1D)URJ#1' !6P%XPvEb{ߝy;Juh̘21": R؆P@RMsL$l5zytUd:|k,X#{blc垾(zq۪SѠi49JQ!BKq$B #$ϨnSVذgS >MY[ ^)kx[mxN*+l1W&치mcv*kP 5W\ cԳBWJȳ1W.gS功B-\BsϪٻ6r$WI_ C 0{dv?_m]dI#I 6͟;Ak,׍ǭ|ymaW n##C4$ \L6 ΁ i$&>rDI3 햵#y 2Fˌ NϝV;eLEtAAl85q6#\R=%V:n ş_Ϥ5<>_z] r #7a^Z))O!GDp2Y-HGuS$N-ofQSi& )R(s^%)caϫP!($`IԨI&Ix 2W2/W\_7_nʤ!,~YMz?M%'r>!n?7WIser7nP1@yǬt~Ƥc$bdF*Iiyb3{] GfjzMg[7]|< L()ߡ!8+cm S+$ah'509[(߻hG_՜/N3XWks6j+?Wş~{RԶK;:NL)׾cfY!ɤ4mr#Jf)6i-?9쭐 }ٻ٬|㩼j6vS{{<>?4_ucImhEٛ>\{/`Yal;bO8OyVZʚq<<y!SZXZFS=1Δ _ZUH̆m5 倞)ČQ`a`j }gG)At^"øt!N==žAp`Dg^yR6fFJdWHBg {~(&z6 4[.gD|ԯ?}QcY6`Ync7̂$R+1"!y_M3n`RhoYqxOh$Dt4m 5A C@BiךY V좼 z򪏗JjKOJ)Rdu%Ih[j'|iSUt]i 7;c'0{N`Zb'm'"F1pRNhe408F$m^#FЭju3ƈ4NS+4sr)K$MFɖ7C&ΆG}'Sjw XW"1uBzJ'!gJ-XmTdM4 GWAتFN|Jd H贲$h_y#wh$/Όa"ڷli-ë|:9dKU 6`:HV"ޚ]Q'ULV,~Oʁp2W톭w?ͻ=G GϢi-)}1&F$ow>LNWZ OehZs$糤!e=k 8l :%n+ RmBj[g6|ְ5 e,  MgpXmlx`r=yJ{w؄J^? dhs69aHA@X>6Ί8 :"R`8 9سp˰ !I$1X81n"72Eˈݚ8#J/ݚvlڼejw}O\ |V0 ZpEjƤ j&cB 6[$>r3hQ!քHdi/q^H"AuTֶ0VAD}Ac[D-#C{ V RV^#7\,Q/F *x A*" s=c&^7I* Aq2θQFID=FҌ+rGEĩvrf\lMKE2.;\:-5U‰VbN8@ׇ%0BҲŧ}ִc[w@n]׿Hou[Up!&jiiygLDe0Q"u+mH@_^/#}0 ۻlgCb}ʌ)!_g"uC&m骮U3Kd7k幏h6+77ڑ%Wis6Ukpm%mb'pG9'0+VG_EVv--%t%XARG"`bQ1тPFрG!eLDΦ범9 f}3גo{{0e8٫Plc KˁK"0|}~0{*P4luFW{֯q.l{OJ0dTfj Gk;r,ݾ ώLFyv~n]ܛp>~$ )5s‘ƹrBПZ3z&{M[x@BY0F(=J 9j⼗ 5k!n-\q4$32IG݋]{ٕm3tL'6њ73 YcKPOaMa2ʨDY [u NcF1ZleF͝QFG*AJ$;A ,PӠ`~jg=u< &qIc!3E \ў,wt$y'AˍN+8\gX(BZhYB$JE->ʾ{~_`vf A_'u%c|?̚1@ӏ/1Rr7aP%]_ðr5a=] 0 c9Y藺(K%j1ǯ0O^oU)%4˩? Vi=VJ[úVͷ#=>J։RkEl#j?RHu*->!#W OJW {e I.![u[sh]jWwjwkWf!̡ #~F;OZ&5z^hZҭfOW߹*veyVrjϯrvqh^tꪜY:C^8^86;S~y6i҈OtI`E"ʹ*؜pT ,X<;>F.|3їx2EU=Տ/6va+Β"7R- 2(PFlpuV;sx8uC>(3twBVhiyy2ia9vd~VdeeɫlfO|g2ǥ_~JcZTj/v8i̾ KRU6a=[oz,nmlƈ@q}G G/'Lu7gIc΢E*ǂJ) B['lϝdll/S[,)-Gi qI)&LZlrRNq,z̝vX- B7zf!% 3Bs6/ 95JO|z?|8s -U,|ɣK̉7#}H7[7Vn,ڍuՀ)?֟#'R6kHu7$_|4Ƶ ;nrM>LȉBTmSXx=7QP-QF&0aj5+&VYO b{43$ւ4r:IkVĎ}jl|ڙb=7.#Sh,C+xD>>5`lbg@;+뙓\)sо2dO[c^hk.lDZn=)%Rx7cJ`| Ud->6aMGkW+qgVV* &NHo8S2D8A6*Ա,*t;V#lE#$YĸU b؁b c6*%Y@)a4 JI8Hq=*195x+suzgl^jB\5<:É1 e0n^4"HE028pbd0XW[OD/+5OQ$EIhv1HN"Ep9R1c'RG,!DWVVur9~8;q#RzwxrO@'~h|>B)$7+.s8;7}+\Fס_dYe*x,#E(.LFcku_]BZCk:3tEwP?oe=JTK%%ٰeb[M9tIc,/f5"~u(|JAWuRr#&W,"۪K:bdNqeUG)p|Izrjno,e \R vC^.+ +De_6fw\1u%vH]yzSכS`XvkvD#1iwB1ϢY?t80Α 2N~2&7襑_Ic0\ǗEwxP 쨎v"=4t"?<57en?Qf?a:Jd";N>{,5Z+hΤƹ6B [4h$ONb:J򒑞S钧?`ɣB|͵$L&+;tNZTjZbUL 0e7ś48:;}]TX7ʏ;h2. Fn/WӛK6ÿ9@Yǒ8irrNk(!U܌2(k099aINSBOJ/ ط({쏧|M7T}WH'tM!Kd.C',IQ%p%q6Ch/o&p>|7~!_C4i: &YP6CmAH=XI[A?oP@^F &eP6DA'*J٠ߨbU~ w~z;po^\#XkZ^oJ9[ڥـ'/M PM(:] ,\eHj@Anw' P`ǫ%c&j{ 2۪k䐬D0?*+F]%jwuTUWOP]q&Ja0Qɾ+2QתӘRW]%r8uPIjTW2]dzRW+ 3u(  TJ[uՕbq+x(q6ώ̰[ܵ~t#|R;+ ~HmB$r[u^z! >MTk S&_ô'Um~(FNUFn&Up%Tà?1u{OhXH&J{)c) BbIA&T,/gBw]N!0xpA4_eoaQKR[172zlIRd/(m$K\YREgs~6 (X̵Q*g er+gI1ikR>%%Ul>mT,~7k0%P#P:O_,JԱ>!#C(UFcP0vYwm#I_!fnA岳fwd觭,i' j%Yh;bX쮪UH  `g)Oc.*5X@Uie R͚.5aRiU9UxsO6>e֦Eͱ/ݗ_ʎ_f  ȯ*/:`1t2l$i/9XQGH(򹖜N)IfxG4n2K]RxcZx֧%ؽO=fe*wX5^)}amYI5aT*̿E^m~&suK_͇h|;)xU{KV6rk@cQaüM5Iƃ탤ÖKjGP6nz!HNI>%Ls6'Ü9Q [cI"kC@=唉>)}S&lLR,oT[Y@" a8hV` l9G`@<0 jr 8M(rnw#"v;UjLH!mWakçLV+hWH{*JyUp>GBK7ጿb(0G%vKғ|H\Z7'b\w \U_ky0E2okR#.'`r08W\p*vxm?O6ܚ̓?^]KҾC+A1BGv.jϥRŠiF΃r8%C:K*Ljd<)VIM-a$EV 1iVV7e9QG=Sw Yc_K͔Na2ʨDY [jU'ۗLRjҡ}ݹ߆ *Pk[盿o^kזo^U/0Ίq&ׁesv8`j0쇩3[1^}!_?^Y R܍0I\FHZ딀79^&Z &IF!lUH}҉ORڀL$22 TK)5X1"zbs4h"PgH3Cb`]j]+u` "&duĀƙ77lvgeࣃ/15_9pFvx΀-v(WJјV:F靋'Ԉj= 㖉:}Ms>配\5R!2ES@ ߢP0=! "|SlJ_A튥~ &PVX6< DkΡ !&N )"I (`ƃi:E6ގsYX$˕w3 /`ǘc|@[w^PJ"A:P8ODjd =X1- l;HvS+l8b\QnF753ot?X OG猴"2Ǩ&$;&A ly&46j-0s3<t~;Skz:Y& x@Ga862P%ie>s+# cl,O+'uI/Jw=q–k=Sz**)TT!j뤶Q#3QIk)QV`i}kZw)}nBm? bAmn`3zldKx #t\Rqr3N 1 e0n^pLIFgC- k7b['(FQ9j4%8I8 o$:Z '#3x"utN.hή:;9CMYH|=;xz{,>K1gѸxhR՚_G.9,źLdviF~X`>f¨(C7W-W2RL wG cu ey취&˚.P/3oS1Wfu(JԈ#cI':>r:i E›-x8ph@e0 )(pFɍZhԗ|<Ed[u#wW4ZWR`y I37iׅB4k:r{=pd/[5/DLmծ/_.f)ft<)_:bˉLATg03I #Lg lv<=韌G>L(O PͮHIf|twae& `K,A2\L6ŨGਾ?"=mGů_Men_Q7 /4M?2W~!5Z+hΤƹ6B [4h,|~4a;_Ӛ?oPsÇ_Е&)I(!Wv 6EOSi Y)iz0JA0)?aA'i|1pfxpT qhҿ'|V^O/fe9XcI̴tR 9'5Ő 2Fx51A.Sʈ.XXB\$^ & H-`v^@/Y/&O~&' zN4 ? 9Zą'o^qU-G)UԸ_`+B5 rmtU6恅TRCD)r/{lJj/VJ3`jjszU)~u6^j΋QE51j 3ꄷېݖ䕋Kb80Ӥ->@۹0/€+do]'*zw~Q>xPqg"(Z],5}T-mkK*]{ԴFSB\镄 jgl"뽹E`ݞĔ^NYazk3x6el 狁^?n{&̍2r{lYARu1kNToM(2l=5 Z[8'4H=t ỷ?~|sϟ~{Cy{q#0{m"MEap~+E\]k^N#{}(qor˭97CsT"vT-w͋ɽJI=?oZZ*# [JOCsx>僘GA}by*.GCF$/\I" "DX'2g>xRbRȃFmt)ImmX`^U/ohqHNO$,'}IHriR3[m5Fhb>KC":;aZlNlaJISٜ(<_y|Y("/<99`6u9z.r5憻/~ո'[^ʹ| "'g%&غ.r!'EV0qWo|?äl.c&%Z2IP{U羗Uqk]^4H:j' o#FBv zf; rX0ޚ.ذH|)nya |^Sh)PR%gx)QihhwEض9/x?ⷑ,/0\$ȩI$ !+OG}H \^k\ž$P!(Jblu%e윒bԚq8 Tha8w kиĿnM0xN "N js` ӌ8NYy$PIHA c=u K+[qK9CxD+M_8X=:rPZֵ GbS"gcN3Ldm2,D]QRnvFRI$4Q;!g$2xPAK4b3WRnˑȾc~Oֶ*n`[CA^(A7Ro:|ݕNxD V9Qoш埼ܹUlŦrd |mK"Vsi!FDxb4 L%E(uk(edTv5ʤGLv+fwc|Q>ͧ@qN+Dž,V:HLZ;F"jVH%̼Q;ϑ98 *E!@Q*I>yQgBi&ƀ(]9C"e`TH4Mn@-YoV%<& &_O/o60 :P!=oypy"L)EV[%g:Y gX eF pJ\ HBS"A|\+B>оB-& 5MHKrNM] .6O6-vp69H!^y}a^8'gu؛[_t1['_.U⊢ѳE`ZKJ4 Gib$@"r(9WɉJ+Ӳa RBrd:$NBOkB2N l8!TMR8k*Ű ya,n g}̸儻Xv❞Ol0m/t#6 #9T0$h9a삑Ɋ :"REl0Aʞ pg692$ NBP̠F‰`d1ˆ]LۍaE.6;6Em(Сv`op]\ P4fL]PyV6ϙ _ !#3ZQ!քHdi/8Ɯ^H"JuTb܍R9+Xl~l0";D5)y4ހBY8h9Q/F *z A[ 9o>Q&IEyPReTGMVQ0"g ㈯!.N%[gY).¸(;\pրꔷ\8i @24 'H}CKh@ IfǦx a|75=iGUqw8y" ٸQNk8uޏ%ŚޏbP%ZKۣ!q_ԁog;k0DJO$s*FVƆThBg[.O州}Kza.ȖH' 4q4o$с>ipT*v]K0iĺey`s1'+mǴncZb{LJq5:V+tYH0V LPR*WbRn/* *0+gB*"4;U>|UM5qWyo x mK0`2`U=0"ALjMu-`2Qf$VHsvO{S׶Wp?Fve+M3A:S*Fww!L=Q-gj&IY.2)=N*ۏI)-o_T'RVkK`-+OR~*4'!UOܣ֟jm$JgW(> pTWHPW'h~ S|χh??擳Eµvuo<[=ˣW~3W$JDy \%idpqpTUsXJ&qjgs?(ƬPsDF6x9-W. Lw ~6\ۿgz0x*ᅭ SBk%\-Rs-v(ۉb I@R GӠ \` FÔ2 ,Kl.e 빱.TR@e/p_Q9?S<>PLM؈%ޥTb -9KU&nn>327ﰛ7Ӻq8N>:sӘQcy}|)@6ŞTBU9&GudMsd>wzvvљؼlܼZyJZͿWtZ]|+uZo\6h]Ӻ7kf mZ->ۚr&lmpX/F?618X'Ox#NGZ/;Z)$v4ǵb6 47$7[Z7nmؼ.'nXѮܒn/=w1^~y%L'gD)PzqxǏ_Ԏ~@OpNqd*HeD/iDoGn49!NawT'qH?slG޵2K #"87"{x}B_e吶S=á(diQ6 8隞2# ?nPm 4al%B)9`Tԣ, ^=}l?~vD:#`ؿ\ ?mM*gA>~k= :jAM%ף`xTMfbh9/ןPVxs]6LNU" <,ܚ@GF=ˁuNtw0>7<~՚byqۤ}|t6Lxr423X`{杲%p&| ;fh Oщ<TJ餵!V ϼ`E[ ,__`ibl?}0LoxGw90,HCH,GQ-,#4yd:({+zC:vQw)ش#-܀9 ;U[ 5/ , jT~ x rI kV$a_RcKzĮ7;רGxY+Oc}MdKm?yӻbRW5sw8 ,(c @QM[%Q T68Kj$y}G]҈;{Iݛ0dpA,n|e]0{̠EmULhk~ ހ٦zOnWۈj޷waG?Xosoaf>G֘KCBќ0 W\3 @Y)JTQ]/E\&2F5N 3x%LZ˼G놦h4 1#玶ӂwhKWPM1}͛l]|>t5 %̠2*86rJEåLӁr sv똓qEBnUB_u`U =U L+BKІit҆ 'Bk4% (.hn >W[BƝG#f j'Dcp$Gȴ7M4b srւx"`(K-Mw$`*땵D E|M PcE@e5J"eTXmTV6@;=(OH* !RG*WQ:7 Ez8 Yoew]jd^m\Uikoojs{oH s{N_|9\RAhVUraWm'>wPYײr?`.y8o+d5Oǯ5hd^PVS\4Sz~m'R@x(^ jkUZ,ȍ8opeG'm<9Rf1?] J8CZ/;8\ "c{5^BGGӳjV\*Ngߚ9RF6fN&WoK4?N.Fu;=l6};bA[QNvs.9xe N#-&whFzI#е,3+[qO8]b`zu9+GeQ\7{)m2#m< ҳq2lZ݇_G\(*rRh>\[d?!gǿ|g߽:?ǯ^>O8S(#aIB6%_#@߃kCwz 94͍Іe\]׌{]|5fay΍l_ pv秭7X5"d+(VJջ*/q-y*TBaoFLAP ^/n.k-Q4E|F9#0 IAC\U%@Ax4;KFҗ6Wi=;8YDoþHR0I%EDr+\O 9)H :)S #xbÄJ& l| |GymӕΑL+}/eMn'~R/xBM2:yk՞*[>wPkh*|B 翺H7z,pQ!W+v|D5ӑ'έ!e0F_V F'H kXCC[/,NIIN'F(:t-#v7J;'@u\6&HjfBJ4Á(K12  *@*v]%Igɸ%\AAIJ$1IϢWpeg+4ԫ*# @jsQ#!ÌFc7Y%qV}I*7]W#s+B$Bx)e!*'EAlZ t{`{dmn[m-A{wB,܁y2;C[P7Փ]%y_ OקO:^>ܩk^NE.#'!X}4[`&p?r-_܋Z\ʶKm+w?* LQ.ZHYC@o]/>=9Kdc')8P_L-sXVى6&kwYi{9Õ2uXmD!98DڨyRflvvCվ$rꠈnnWo_qח[8KCjػO3(ҪrV1`*oL2pON%m8FoD\S,*p|/i~@B"QY2  j1rnT8-P-I-̳]I.'vc;ұo4Rքtv'm^fnMk',M_!022(1u<@#1Dۀ >Z}$2#4h JEe)=oeLL2&cEѤ$\祸T J#c1rXLW)8cS,䅱pX(t55; x{AsބY? f#67 e&/-7\NXr!`de+y"UEl0 lqaW"jf"FmĄaU Ea.Fǣdb jMQ[FmG{cuf>RO%r1q# m5kƫsF8SP0CQ $˛<҈Q{\. MPCn Ϋ՜7S8m;u}m.@яȬbzk.qNF8Ń,`'0?FJT ^]s}-Uu19ХzU%\fŷ}Kz i.-9QU-X =҈6*b"AdTSMJi.e\ְWRV 9YZh/__liZUd^ϫFPkaVJP'yi+1eж ^{yJ%E׻{pO9rSNߵflK)'R>$d6rGwem$ mK~ÆkGξ8Ǝ:%x Q8AD-I$&fF2Yݵ(JU !p/[MFt"ŏD JH*! )Ҷ" d#x+3#^LGNz|/!ڽ8-9d Ht7~ZCyq;QnE)"ȨPI‹RU02Ar4'&Y1%Rۭ,@] bWg^.4t]6kj>gij?r/]5SF>[yc_4ʆ R&e7ƫLTmjB YH\~_ޯ )ں S(T gV0iyQuURܔa? حWR;t)F4ZVtuR9AzN8N_.+8uUavK^ in;-U\wK}H)|Eqc<ׂ} L8B䔙&nXVz6A)Da )ŅG|N(b+u`;km,"Ѧ`Pl__B(.v[Ɓ77Voi+'#y!S}.|,0NtV~|l&ELA8nyc6#vAoۻ> b(ԮML4ZkWd<^[2j*B~5g1!lE/bJ?njwc>Ln}[: S?O>N鵵vx~h$e1AP=ifL> iAAR,Es$A"U[6:J;9}mv*z] bk 1BSȆ.88=x-z~<]Owmq]isin^qd/,97ocm#4#`7 )PL[e9:'|bIPČdT'sAQ4G!P:9,!~;m&4I' KT-+m&kr-䑤?I='bH:I KV4+"yt1 x&_Wpsĥ[VaoKyH\ENC !U.$RjO^;%XNz. %NSӸO}ǖ9[}f*WOQpHo|Nz5a0Ul29«R1~. ~u1ƫ YӷH?}GL'm{Ro۹F]=)}7ϗg~KMqWxJB6rkdM25%SlVnMa">?yDiMLpKKϫ:~P]dF3l}]8=Ȼwknvj>fHbHH}9AzaNzIZѻQk-J~}ڞow3?媞eeܫTϗvj@jF?ſAAY'.N)cH3#~OK~vy>m-;$Iq0INJyn9iy*cy7XDHPRFH&\A\PB{ƈczJKZfɕ[,-ocaRh,m㼠&Y,oxs֪l\ N̵-=Ty?6hmύ"j=4l˰DI]J3pbsAg}H~3jE0}~ \t3D~rUf7礮ŕ=|<^h&K+juㇿ"~gHuwmm-{:I]%|zӛ݃QU`V\{(XJkپcVJwboiO?-M qug-*ϛ̩4k )BImI+\gU *k>O#ڷdGGM'u#̵.tڧh79[-&WMY3wUvMY{]nM3 Ֆ| @Q$SLDFM$g\+?=+VF:k\y{:?h׽Znj]mA|o^5 CGunn[x[yYLϓ:^ыӆ6r 4џdh~sЁ;@>mwwaЁ/Uѿ7p`_k?59uFG:B Jrbi8YMk|oU:|DgYﻗ^ .{Aʖ̵+D:T$jgyԪh빎K [O9-۬e]6$ajf}+̜sqO?-˔%E<2xQ% LWR$c)I\9'bivzM8| ndáB'&5J(CH"&)j8Òqqۥl g땽ipCE\XA09d Mr'(DtJKQXCJ;DBV:'{ B| DeK(SRPEe̕ @R*r*s|%a>MX YT!RRZY=&]X41#Zuʩc<tA0V]#lbL&`LxC(*D3-"mm RҧR)hY0\yX˸3N #Dآu0scоOyr}c?df$FZXy@t+mC uXxbL$֮Wr(Ebs,#DfY[xi W^0QaJpgC4Z'T85Lb $`I$X\,RePpC2ɱDŸrмeK[C0NHyE"g\%"f0hX>u ]A5/)D ;2X"I]Ry!dy|)VNv kiGC K& 0<8&$2h?RiUfP3@ksv[U:TUYEBAX1y6|0t ][:h4 %xCV$+s ϐBuF\F V`j]iu;QL$amV#TB`u|(E BI;ț0*!E]Κ!-d*$*`|VEHFYD6  .uFq7VGr0( "g=/2i Pf6W,8rfL:岾QrXD/K@B]N*Ё'BnTz98|,B !RX ! eE`H]! B* y@6[CSX,&.\"0E!LL&FW LK@d*JA 2tk+Y {S4 1 J8 92r\`50CjAvRP6ZO)#@N)!Uˮt@*6ХUc܃.9&n$:?gMeEyelO#BX`[$ x?$-iR)?͇hKE"OW:unuLV5zMB&Ȉ0%Dxu)}|R{ȋ9i"!B.-$pjky $$LAR{ XRr!ns)!5ڵIҎaq r.!z-b@ l/5&ݩf EmS;"hokcМ)M@G!ͱk,F)I@̔dȂ&Jkcd@@q@!#2Ϊ)J ;a!ԲI }ڈV`A2֥9mZu[ 4^Fۙ:TgC& h%A r"-[TjJ1譈WGd~ɍ`!-m;MwH"tF}\M-i0=)6՛zv3+dyzV5oSpn5@Id,ݫyUB*ih$tnGB$mwK((;\[ huܵmP)kE Q6'$IbjFCm bL@j~5l}˝ʌ],hw)1Xtw1RAA ,k$ hT5Fi SRnEַQaKl+OX1( "NciA(HmL ww:|F4F.L Re cGQ2iCǰG4K` T@ QŠhc@sjom:lVj"U)J[C UCL/Q17@5B5gB'еkiq~sP- Qє?joX 0%6m]k*C-2m2mqT@: Z6 /ڄD\=č#ȍ9>dN;=jM:CX24ₖ4 t A.2 76v:kpUݚn4DXkAj!#!:m.Xz׃*H5Gm(O]gF 3%8F N}KTQ~ԋA?d2[f@QU}S%*Vڢf|nal0o0:nFe~;aJn}-}aA ׫w>^qnں_>x"5zy~e)3vhP+d6?ҧ`DjJ#9zH5ܭwJOb'?Ob'?Ob'?Ob'?Ob'?Obt'>H؉6?C?ikrv6 Ok4K=y61^ifA7$Gtӣl~m']~k.Zj!WΦqnc /δ/4vnTӻm kiw͛*]DͶ ;@u"G^Uo Dr~ڈ}VaJ>BV[W&EWńē -YвHg0[*Va Ulf0[*Va Ulf0[*Va Ulf0[*Va Ulf0[*Va Ulf0[*Va UPƋd&=c\zcZ[%*ȒO}@b> }@b> }@b> }@b> }@b> }@b>C9[|@26>CyG> }@b> }@b> }@b> }@b> }@b> }@bС/>/cZjJw-o֗/_Lyr9=DTO%6?%7%o["mK`[cRv7tEp ]wY%h/`ҏNWj;=.]m6>[]-1]ݷeZ~tu\"BiM]1]}R!jmzDW؆u/tEhw"Q2] ]ݝ蓺"F:/tEhw"!ҕWtug0S(&_&K%o2|,tzwfvq=ӝ|4o`iQc')GYUc[uZ(m)2\}P׃I'ۋ˯K7w۪ϾTja«Bl fmWU o&2E3[|7@ÆI&n3V46My- [S3_~E8eGB ڕwS;~htxIDAYū=QhmBW!Ezu ]#WeྛwwsXg?6˅kAWUO[t[ٲݩr_4&I5?af"кRP۪jE袍=+ZQ+}o pM2K~ (C+/}ZA] n4}+O^_tE(b:@ BЧd}н+7`FtEo$ $]IunZպ_O-B!fo Wj~gՄ&#!jb2ZP#4[I4r0^7{,<@*#.ggFEiz{4U^i򶙻Ϻ$4̛d|y~+\WʢJ-j3q&^1Oӗgma[_h?O9ϟ}n(G 7fV&`ާ9@fǦB*+y+A=$6p5Vh?si]m=-3]ݷtވ H+U_ hNWRI+% GtN폺"F;]J _V +c_ZtE(!ҕQ&F#2 ݟ+kB_І+BҕU>>]Yн+۟+;]LWHWN gDWO2Hp ]2Hj?DZpT;#PaP LWCW`uH7b(ߤjz-[[ħY\6NFr]0kQ5*ژ6{q3Yz1Lggjs>m @t-Q-nk75Y \-UmcMji+Ej[ǪUݒQoƒB+!;E榲9mvo}PE;2*biO[Qiʤ誘J[)cjKVl[ӶjXCޚ*m2>}pn*UzxʵGN& ;џ'+}_G( ?r*ܱñ@m;a#vp#BmtCW[ra LWmzi}+VOnztE(e:@VL[uEpqۡzP:C+mV2+"esLWHW:B ӽ+t_ hNWyC` w\7Եѡ{muK^zVu5 >V1hi_3I=jgޞ:Fh?f$QT;@RO%7tEp ]9ڊW;]Jtut坉{96mq#N;]`Lkns*-JJ,i]P&VQߍ_'Wh M~4wצͨ*Q 5'ZQ?zMoW g>_2 . m2Pғ_5]vftYGs7pDv\mli6TRbںTxZUtܟO䫒 ݨyA"Y]I:O|Ull:+pӗWW016r/xC~?Pu/G'˯'ѣV?Yen>q|77c؇Xn UͿ+YwϾ]/Ϊ$T^5N*u%j`庨$LMTGňhjy *F+iS$WͱNOÑHl:ڷ>HךcY2rUYGCqA2?d8Nkcչˆ }#QDcWN1_}8fD"9}7ctUN'}N'ao"ֹn(u84ȔܤqJQ߫ׯ{xtuҝTX~f:Xp,;ruQvv~9OWttՂo']h܉;+~~_x<#\J&u~єPNMot2y9nx+zzxuQD}jeht:ORc?r'(?(}\ΞJ1fziݺZ dspgkfpݝ"r2:'d2'釤?"17qvocXzHw[$-.;y&Z}nHCx3zGqEjIcUn+_p~e? тҁUM ޜNa8ЁѨe]An%nnf]}g{řƦN>E6 :$]RERMB&SBUClOU!mMjҤVQ i2IA72)"8g{]])W߾}Je7'bͩig''+6>oxWM^Pq7qjq[njz_4/xۖ6kךltTh \С#Ÿ(SctZ&g=nG2o{#" ^IHAB_%(KR ~g%qHYEæVuMOUuUu] np<={*zkpΎhu)jq.z9]q<~97H%'^Cnt8A nźLԊ2]AA.Ԓ]]Kͩa]p9K+J)[-?3dz)~;&q[@W@Wtt?'/7!AoF6Fm =dY|-^[l@!|z{"Duqjrb׻ x,>Vm@`h#Y-dV}m!6UbTZE^/*KUH,헝5+"nzu xv"{пemᨊ,r=l3QA{ ԌZO5pL܌yY|Mj21 Dff ăi[o?8#х|߬*LO=he>< m:P:1:! A)6JᙗL4b!rydKXQtEb@<}ޠw1;J(̘̹`(>ĢqʴL&RzL!^):vFZzZnAtWAns2*86rJEåLӁr s똓qEBHnV/:^(̗;dJ>#%)roM[] %G0OsT]oƏs/y&-U.DhAq4(8 Ax=t0"gZqNy8p \dl~E .ȃ@ J9׭NU_J:Csa(ˏ .F"Ո}3_BGG!5׹AhN9́2:1 s>`|' }RzftQϮ5oCup-AFP~CSozmYH11|q0 m&ZgrLmӰi mfXAN ++Xspq8_zr3g7]pu2zCBQ}TT!|&[W(bbX{v(#Z*_k`9w;.Bot7Gom|w 8BM _~ͧ^|M͍`jû|f7Ww-g-o歌^oqK|z3&!b\Ϣ,|D,_)Tܡq/Ǽ*XXNd.Gj X,;}>M䔤` 4rzbHюY̫QUOvGkw  BIm^8 #^H)^`8e"F&vDE=HWAvS^, W19QI6&gz@ S @%]WD Nj5rIy ӹtIxx0MAtIF%Ѿ$[ծ02*`P0^J 싐p7EQ4{keҞvP5g൧w˿tmk\ ڻ(.Z~"/!ʔ/=g̙Dq_dsc؂Os+6='ۮVppmV%VDxb4AGJ%CZhL g)ef1T+wSex[9p[tXvVp.o*|W|jLgxCsiUVS^keŹp!ZeKpqomwi2@Z.rE_08ipjʔ,o37J]7afMR_*r^YPg]D=$F:ZKo$A0 G:>Um'mJuf/-.>K@ -eNj Ae1 ,-<Dh U[k4Oy**)ws(n܏u[9 )[G"TEqq+ه{׹wb08h ̸^b>=*s3O]k7>՟YulA3/oK>'xM*k4fu vL&w #ȋiPk_J a ci+l`T=U 粋5q~^Idw`fHJJ ɔW/Ci/} v;À0;a`;0xi$n?q8^ip QNNz.AIdHxB)|&-@C8M 18-(K$FEJD3FglRdY>'[^˔0Jx$TƐ= Q512rgV:0%鶗&m. BHd8~4XG&A"19W|^jBnj:UcDa;*st+ ؝ɶ}5َLUNw^_3+;2|o Rs2#_JNq1#PG1 41C$ C.Gr3B6 Wm2:Ř'N E1YϢhRRQ*K(AFbgJg3Ve!'sMfS߬N-L'aڟ/~=tc7~\ y&/-7\NXr!Gee+y"Ut*q$p"#{Y5$%Q31vh '&,RDPXt,;CgA\01wEjwFJmѱ{`WQmFpXsT77е}Gigs0I!*ιYRm b|ON!vW%k9R% dDB-ME$wItOZ#U!J]k+oɴ{y9"ci 9Y9!#$yzY=Ω\DN."$JDy \%!5^8p8UDo?_9v#ճPL- !92x&a [ **$TF a`E!I/ÜKjOАb0Q &L,̳b!Mp;1 %[iTD NdWGȟ,x#+5W9XEt]NNONT2y HDAs&] A v%n!(\Xύ$!\nM*$eAAZĜS)r꾶#%SE:S&)&lx5ƙ9wZql5:93,= Lԇ"tȭ]`z?>6~I?I 4H)C"bRkZiX%UD$Rh+ LN.09;(De9ʸJJizǢJMJk )j_YI[2b :0Qv*ķͮovT=97FjSǥ3e&f&8㜵J/ 놶_Uȇܙ,VI뿢Bx_xS|R}~O;&aWtQ}E`Nڻ? pu_oFq7q<`oI~j(gຨ%d[:7&\Kyջ#qE|7%qA3]E[ >{jN>&6>Nf.5hЮOGok'NS:A f5{i7;Q܀~AO?tò-$]8;;l|܌EˣΔϛ?z* l \ei9vRrRS8!B˓,.S,c,^!\qe>%vOp2`Vc,R^!\fL+;Jr*pGϮPJ ++2(`ppUT ϮTzp%R\I:v2NPZG3R ]_#\)!CX!CWS+V~pP•fȇtXlY q:izλ9;Ƭiaw|oGƠCą"_y]g44Efg(}gFH~x{vsInn&^PRO|7!M?iXΠȸ0H.aP <!uj棪 y ri죴,/vy@>毛.bO_> X;⿦63@qNb܎+{; $Rq]1!U9M3ʹ2;Hs?"E*RHxb*'y`NE(+(w.MG4M 6%g@Pr%g@Pr/OKc,9JP3D~,9J΀3 (9J΀3 (9ff]UEs,9Xr%g@Pr%g@PrE7?nn!\RWSq];K(Kuj%I9S,)NJ⤤8))NJ⤤89N *6+!.'˕ Uj]Qm$"*NYaCCу ê$I[" hp H?W*BLyIEv#7B.d- 1PKwQ߼3&HG铖GU:ZjU[t9"F^wQ_̮2 6R_,WQ׏6,}Jgvs,н- L6 OPH$ܥ>ߞAM}]wrչ{t:\t~kR0haƍBgX |(|GƖq۷3"S?wq ~7u=u?.]_\܂{t;Aй?:oc'8.w,fl~_w៣]Y? NX__=M 7+bYƼO,=/{8PsgT@Rs#:r+-(4mQ7,=W60d$`*\$(sIJWS/GoF2œJ )΍ sK5(B5C,QQH-$ @ X"XNrB! 'j+UI:i-/\}qZ?f>&b>T)?pW]=Fw٤"{RiGEDJk*D_唷HU+%gGw8ӛ\p4hz"-&UFG 4at{3YOE(8$,*%% 4S)`%MxYƈn`kvټ::dѓzy5zZT9cuJ ,y)C5DDtIA ipg6E ^!4.PJUL dZ9B$, |AV*H\ X?1438892\w*T Yַ}ך8^^9_vL_mtߌaYjTdeUjml*Meew%52B z]en`"&QFlEZH!&:Ux箶l>^R4W$V ϗ (l^o^fbSF_14WDI*RYLF7w&ekfr#ڹ\5_޽{7l3:vG;d#twS{cwxA9c̙ Ņc F~S&uwXW.r2^ⷉcM>1uc/𑠔1 yԅ_|n;Fدfbf?~ivsG8u|]wmIgo%8:-pXC3E*|x-ݯz(K#Т1lZNTUbϟ*ԖW)U!6(!H$:G'ob+br^RL[@A˩_ۓɿܔM~d@՘`#ztl~f9-iGY \Ii~0+S]&d:5'&ӓܗh֯/X$&Zi*cVB;`9{:rpqAO$`wbZɇ"uesVPEFx-lI oS+#7 *ڃ~cͰ+M:[˻<_c6B\Svte~v뚲U%uK}6[7(ӿRWgI:h]+5˧Ÿ0`WliN`AߪcWaBO%.J|gC2B6ׯkʞtw6--/3I˾.f$-U- oFwy_M=sJv.0'Nxp!x5Wgw+m0Y%<$t*it bP!ŤZf%zQh0>PzK嗗XQvEb_pfix u7Ex=7_ԥ>)څUOGRqB_P jX.Pu^P㪁inV{]E1 RRCCVUԊ#w1%5SRq_pO#ؤI.J?:O~^6Ӓ!ءU-ǛomdpMS;v낏5\b(1[AK+a.%BPL삏oG'Al,6m2h{\Iڈ`ԓi#N҉FzqR)ݷRڈKݫ̨&U}/[V~_R&,q_LAAJq&4}]^,eV\̐1O!mԨy0eG奤l*S}A7iɁՆߵc+䫥58_'\]|3"FҪ*S^k1^$TYu!Zeςem~1;uޥtR]>MkYA޼|`pgkVx-m##٧c>ſ9ۛC1ʳ7 9Y|:ptdNhU+ V^z[e!(NwtΪt{vHOaY%ZZJFbDt'D23DV+  ӗn/j{vvLco>Bf9_];(^Ӷim{SL=-: Ph"*@$+Aɘ8t&;,gdnkKȴ*I@ddnRC?D!'!!|` re, &t,s6E*)iZ1i3_\d-n@ˋɰ J46_Y'u[oyouNي['n*Yҝ$R)(Q$lYY,mT1 xF31fYx+cb.zR0H"Mt+pKt@kd쌜;[3,3vB ½ZP/eܲݜÀM]|4kʍF،, ~@HmҲh  Dv`c+("s`2"6 8{.  6,$^a’GLYMhU)33rGl? PP3c{ β$¢dIH@ $Kd=gӥ2E^hpŌBAə Zs@!xczg2{*դ|ԖhY팜-hz@\쾝yMd*R)4bRˢ^K ^7MSycAk^?bAJg}m\MȨ 60rJECLFOuI@1r=2Gu:P-[X~.מ^Њr@mƠ6fcQk4$"9@ JAܧcPĸˎw7YzuX@mN/K6ɜh!,ΥKy"gGa8T6d’m)H9'.E"i&сrQFOl RS` ŔzCLCb<(q̏`9Ǎ .︩"Yi^N7b?~ۗߕߜ_^~S.zq7qD30OԿh#aCBv%@M㭆[ m˧^N6.eܛո3; k[ ߌY({zţ,\iM*p= Mb~6aykbbMU85,@8n /X߿6;'H`~>CIv1ĤChUF¤9 }h6.y,}lXA^UKχC 8("1xeeK>f@=Fj|Yht:Y9Ď Lu2v]9rkG]xn1I@'m*R-oxrDY srVzyR>}"OBB*ru`& re)e4y-t `=1gB(1B8Te$S!1'"ĤFwd:|V貌%>';ա|Pd0'$ĸc7 !lLedl%{YXnQۃA}8V齷vUq40[7mҼz"^alZYU|Q-3[OOUU%c8 ˦ fuQOdZ2IM0doVM:Vu[!>MsԆ-DzDWHOd68]--̭ɽһBmj(_xV>֬K?{EG^Eҡdz {'ՄdO+bF8?}è77́Q6Ҝ䢦L1 (e-gF|Gmp<;igY't6K@/O_,:w./n_:fsWڗE KzPb47?Y.L`VgM|1=/^}?hz]/L7k?X tMtnX6,>ί3-1Wj T#7NFCY(m2Z[ UN*J͗ɅAyع0WFPQy}6gʷ@brp#Rʆ,ury._rU}VC$Phg|L.ٯ(T*Xˮsَ;#gw"ew_tx;,{LILɃ{u]YR\}*#wG+`S[u  /;Aзck&k'߷(Jùp#q1f$LbA2jφX9iʨb~ ReRbtddaG2ڻ#8rev=;>( 6O~$7XQy9覗֨T%\& 2dvWM .[;ixp|砿'\7\Fxnh6DFvXڢ2Nbu.?/p=zew9A 0rJEQ"Q"'r%ȉ0;c"Ѯ'\A~v3SԚϙv՗+c弟qM80ɵޅiZ#UNl[=Wzc\Tu+g؞`_Z+Qkq%*-͸:B\iٓWu+\\W6qԌ#ĕ6PO GO2(rWJTcĕI>WW"^p%j?v%*g\!qߎ]]/ D]\zuy!F).%+ʛ˫ :NXo7_|qũL~b/ vt=|p9J]6>]_ϸ֪r]n!_-v%~ :V7_o?{'eVn7/>kgh)E W!B woN+2ǭeO" ]>X&~ӥYwWg7b#W5@Jj@GiɋWl aCV9M] nu:Y|wK ?;v뻳7]_{ikG{-olZ&{ 8vA>ԚJ?gy}9:2ae]7!+%SǕ4a9`:SGDW6L>J恰cUoS>ԂBOMN0qrÁQjhz8zbϼ\WmzWW"^p%jqRg\!tȎp%W!q%jq%*qu2q)C7z q긂ʨ܌#ĕASOAw+5+QyvW^);] ]\M2(j?OǕڹ3W4sOC=.v:3u\ʍ3WF D-JT0q5mLd4jSNV8q67fFO,2>츂fUݼÃG5ryT7aW9qEóvtJ+S&'C?S!T7D39QY_N&gwlz{b:z{}9k2N?cq*V3XYDqr#+%rSǕ<q)2sG\g]P$\Amt>v6~r}mig];>_#}'6UZ݌rBuUx0V+ݾղ i8|D|>/|1>^@]ۛGʍT|H %cV,ww}_o֒4޳2í=W퇳kt'WnDO(jyvQ/wrmzHT4m Cf-#*䌮,uώrƒZY.Fs=f:l)4jLBJ2^i.Jv`U}*)ڎ¶& H}gv0j TDH2XآK: *grՂ2' [H`>rk5X*@V24F$kiF ҐS4)9kU-6kի708WjŻSDmՃkLdDRMM!F$LSm;Cwpf4㠵'X3BQ7&= 5#CXlQ-ġ䔔^^{,`6Z=oִTmbJQ(62VX4 CȥUΤ: N=]Qh~BȂD2q;xk]ѪPBtFޚf!VYpPMbC`pbRtpȤ`Fv zN b#o6g2q p񒬴!5BD.SFUtfb;J= eC?BZGSmf$[eW>+fVa+1j{Q( 2z0-䮬E&φ5P":R()6!,@/k 1XYQƌ>dȏUBk0=A<+\[ TUE1lDdIƘbccU$;!"/~J r]-eXs^z GPVyH0p 6C< o x@ppi;[;@EVLD!]I!T*#"6,qLE<:;>uU,R "=h5':\dO=.2޵q,ٿBf~?E bK6E?-^ӤBRrE %cI*"sfX\8Lt~Z7ph{ ȋ ":t@2GM8ڂ VgESY_~(}YQ"m dNI$ 黍.}vkg ~^n;< ں! EkXkp`Hc#. s^ s@ MހwK |Y] P1B(5m2 hwp!YBGm` @;k1-xW3K:vwhs^u@8chdr[&-K X1sɺ6mϙhbܕaiٴ\\ӘIM%"X 覫2h&x ѹa0`ؾ˿_h ,E4Ye I{y.j #e`Ѱ9q=)Ǔ_>#|AaA]M x XRcF6JtclrM,B2!S : # T", wB)Wz M҈ +Ip xrp$rʶsA/: ;bC!DE HbQ<TCz i,27hâHYX2A )b8e+C`3Τ`~ "i0XfP 9$bUR盉`ebBv"JɦO8$"Ȱ;tbɡ.klץ5];g{6RVfZgq6[J0V]t NjG,9rۭ\ցGw0u6|B-0Ucћ R F ǣuv IXl~PЍs.O/NH̻ xlH' ?Gmoe叜-Hb{2O}[Gi揷_K7,V*y1~娯aYi~k +\'=X^ݿzċoE#i?쉟L%ւ>l=|Ȑ: ID:j }5:L9逾Ґt@"HD: t@"HD: t@"HD: t@"HD: t@"HD: @4~F v8: `: Hu@t@"HD: t@"HD: t@"HD: t@"HD: t@"HD: t@"HRu@-C`cGst^^J&HD: t@"HD: t@"HD: t@"HD: t@"HD: t@"HD: t@/U.CY˝? u@8D7"HD: t@"HD: t@"HD: t@"HD: t@"HD: t@"HD: zl; cYe~]N ;QM_!A4dK[g/[ز`dK?eM [v걘U3Njz~( JURp W/0\I0R/(\`'\\&\5kyYi/1\)=0X>*\5s<˶Y+sWJ))\p97Pjj暃iӬϞ 6+pÕH~@CX3C WM=lVR,7JlХfs`f򹇫f%'2UeLjZ}(Y}UXi2p%jkAZsKZ1ڕ~*KV3/.ބ1}117zO?Ќ՜\~w8XV+c l/}OG0qWNgFq Ϧ5dU#د(|ͦؠ8We&pM%ㇲӢ!{ϺW_3E3 ۦɇyi]JmI*{!U3\띠u0!ExdsU]f,?ڬ _4]| q)~Z#nºrqjyNuH,dWTѝbVvZekX9W86~ bvpunu麄tnp3<靈ӻn:)}օPr~:j "\VLEVcWE,wKΉom.HY=/_Ůfo0LƟK_-o^~`/X|6Y[0߷=M=Vigxy2:1`ǣ]Rx\AѺw,O17VvmߒbkItwKpq3~:/_cભ&?[?zl2lzRRUMr0\Otq ||D l`Oc8l:Eԯtle߫0x϶k}F]XT-E^\]a\e®76G >M>n_ ? Ga~p( |qt6f7c3Q8B䏓+?jaͺb|^MybzO|ա['ji4Y^aM;;oU 6EƧ#]`]ʼS>@?sS.(:V'X %oHkWD1Ι= rlq 6w3k7CLxc04.(^Z{` fQF5xP+q o5Ol54u nA[`7űMs3*vNiĻ}xi>8f m`nO}85|Y[]D@^__?[NXZɌ Lh;-i2AX.p"!\=OSmPl!xq&kUSE>kQ8o,˨H9[ņECi!f RkIקb}t:9~6wn[+}H g˓1P/Nw`k=PW{:*Jj5\!Q(ݠ;xg 3|+|6[Ulr}x͍{ͅyWrtg6AEĵ|֥oBh;Nѓ=0KZWcWϺ# `w!)XLs|̑s|Idx6)dI1TD2U$xB9aq:wf=*6q.nǂi:xܝ~FBgY}ԫc[횱y<ǘ C.1j;t޽kbޗe; -^ Na7ptrNw zq~&3#;٭[#_|B{u~hoxw_Թ}|Ln=׃1\^-wͺ\s3͡w_'[ۇ1c焸 ܝWѡ[̍qv/T {mzҷR>?!H ɲ\;UTs"J{H7"ꙓtHzF$]= dհ*rd'6s)*?Ŋ\V֞)MI`Mq9t0p~[~Vs 7倏 SO!`Spc^}"*ы`:|4sOkgYi!K5'B]S|iT|pS.hVN+ 52X22h2U&+lʤ*f͘fAؤeֲZ{=Wsfy](&{x]w1#[,[7o츯nVsga7ЙLD枈I/7t2`:S ўH{[{"tTξ 5(+ iqAz#lU %KSq'D`*fqWAԢ +;5.vAⱡZK[_) >b>)78ĜOՎcs$}|l\ysv63 ϯȅEigI&  sQP}zWt̼Td|7-ZJp\J"(a5lXw$wk?&C->`ksBk-C׻-u_€S*&j pV8_SP'x5:?[%Eo/ݘ' "sU `<))EռX'RYAJrF,c|rsBcsJatw[1Ur}A񐆾79KsIgvK)=_ɧ{(gpq^tEM=.8?O5הJvEPIäLYσR乤ĭ3L͗'BC{ l Fԝ=%Dd/ij/.~ ӧ_raB[ZZT) `#K57ig9esu~o珨z6F3 Wq 6—`,?'P!.' /W͓< l<͋mj&nOaxVag| jbY֡4&R4xI\xO+Y iJՑ3f(ݝ}Z6oz>|֢&q,6bzռ41%ࡱ7lz0Q_E=o4o?Cxdf38^E+B"$:^+_M n([`/[b&r#m ] >Lh_~ň, x ^憨ЕtVi5yۼlhK]xSS*(H%/ZD7y׳'2-MRWySSW>'D[¼zAAHqsl,pV?2M;{oήy`I MN&hq#@]¬vLV[$nwqǼ6-,Ѐ>r[b.oqg7tgD&!$(عog2qTЫ_IYl>yt>7.k5Og}!%ZUp?+R.OW_keon~)L}`i 5/]ҪZb.}_n %[]+^7Rqp,~;]1k \\](-ayF >*wo~w}q{.r@ xL(yp+!!Jgb||&y91[80gh ڽ \7\Ƀ(7<8^] H\LR2,=U}.oTRopL\"v{=?;vvm$ 4vܫl5e4ytsi)v2C dCDs͐rhk)I"i%Y4Adh{on}-x4 Y Di􇙒!zK@PΧ ~yz7Qp>VV.xvAo${3x!P{:DYA QQ/ZQ>piԚ:85AxgKPC>1MѢ,+}%I!_$$*2P7H%]*qhC_xCWpeoK\Ac_P~5%s}޼m3蹴*Eך@23Yq.\Vi'A'6FMGxg҃u4ґ~ɇ+G=Y/YnoGo'}4%`4Cd4?V^I|NzH t*R)ޖI`^;:pt}l&^g/O}O!؁ZZJbY [4+e:%G[$hU*wV)WShyߢAe8P3./oqMДo$7{5wK:j2sUsU.v7r^>@GlV8֟A6;K\EW,oY o1m+d3"u.u']&6&޻@nkʮ''mj3W' qlOD O?|0j#Oyl/P.HYWɀ fD}Wc<ݓ,l#g˸#!=Z^sPa6'thX>sMB e&/-7\NXr!gec+y"Uq$p"#{YV-$%Q31v '&,RDPY{#vU\s1wEkڢg-Z{onD)fVx#T,X7RhW6.:gl3 81TڅU CFː & 3A: 8$S ڹaoܮ~"ΠXՈg(q7IsSHFv pa 5_y:\a6D)2n-$*+S`(PTs0!'̈́NSr>n7r,GX/{QR7.^T=E׋{析\zcHƩH6/-ah$%]޴FЋ{Cݳ>Ӈni6l[6'@e x$}GJ+80(Yë?{MBV2Sf8Ҫp=QJ &t7Ɉٝ{EG$(=$2 t{5 %ìK%D1(bL~ap Օ6%9\s'HQęXS*t/<t'2z *-NeG!?n>˅6X'9[Ihc%REj-? 0aMR"f[% e(gJ%,2(Qrt  (l~^E4'[CE+"MNthoʛLʭCtЌwM3A=̊VqT1QAWߪ;C ( \#z,N]MbWUAȉE-|8m@*)Nc>u.]Ԡ#2x?B:8<3?:?g`dŹO8\B;%eP WXJ meur9Y&Rz5.B,&:u΋kHsEڞ.ZܯBlr~Dy]4ׇQ} \)ethS׉ӜQ2=01_iqu^Ckg=9|v8pn@-ܟ*'r{1kW/Mo5#"u$UHBZ^8ðZrubUfX}͎GhX5#f+*q$EEXܢ껫TYE؊pbQH}l$4cs}o]H/ĝ|$੟G0!ю>"4$+ q g\0!RbJ{Ѡmx,H:ذTye#=v1Ed:I*:>Hq,欦#M,D~O=fɮ$0L`Z-}y4xUߠf>m#XkmnVOօG*n7Ik~u$MRrl6f)- %R*[f@L4N}w] ̚o6^xhQ+HZpA0e͒N jAwRiZ|yRl;J wY 9+rV;,j(@[@Tȿ4NC?%D΄bV ySk?q^DP 6 3$Π]rj̸UܡbqF$VN5ej[Fu4M诏;doa_q ÔN]֯KTLx @>3Йe i*qHFhvM˪t͗Id^$3'1":z/z;ufv}"U֨bߝqտ$\p+ZF k(Bx*b]̤ .tg>+tj]@wNXP;ǵI}kwDE;x#=Z}޹p@ v]M! 34F$\REbD2i^(:9uC4|'T43]YJ8SF(1 r2(vQmCy½ig QO7]54 w:{"q糳FCTNjQ6G8sZ%v8iW2 rV;.@DύTY,J1ec$3') 19A *6G[k=oY`OFY^ίHpfa6E`ڮ,',V\oYV\% eɟd-XS OTG9sOeg9Tˑu[#rrF1<.FɔG\2KahwJP > 18-HgՙQY\;[gC˱azVc넑_OkU 7U7U6w%ZѦ6Uy R[3}b [} _J;;*,sI߯ͪB,I䜙vju ]yA %׏bo~wU7hil7nOzMˡ9o9AW\1m:i&+˭͟6WJO}+JDғջ ||m1VD 0.bEgoH7CF:섑XWK3,$^ƌv&C{]>I L3RVttƳ|d2L (~d)+9MmZĹHo.~%MH yz8Q*>w|L:Lm0uYALWje~ D_$@<6Ri'@&C$>G F&-PCb5cjǎ`Q؋Ѥ،2Ϥ̢ϐ5CELQR*:HZ*TZ>'gv݃~0[+lDkv{]ȎbY|7w7\;tDXv6`R*@#dTYX;Z,8ŁYᄕ#2AR1T{x{zר&ݧ>m^jmN[,e0("JMcDR'mr+dMg}p>?}.78/Hu}t'gONd\M hpL>uSΜiE k.dsCq>Y52FhJF&e냶>[rT .sݬom_mw~d]+ P]B}(j۲\ȖRJL*F홫q$Z"U&prj.Y5rC ATZ>z!j|U@£娇bnUxOh6>'ap>->0qpQQ3ןÓ)I|Ĕ}%WGG7L3?6 c Gˆg jo㪒7_ ElF_Y)PD:4x?^WW őV?S,p3W7B*sJEϩgiP~BUE{ǽ  hˈuJ`H6gQI+]qYq'u+nU?AukTE+}K/tIXtKtO4 {8 |'vt]跪%?N N~Ջ?חƣlZMCTdzy|j_EM4:c!䝠9ׇOT9%9ӒQpn۝Y0T7ITf] vTdG6z-Ksqq[LɯS|6X 𭗿zjc킽ѵ,Ͼ~J-/);W<U$3%75ɣaJVJ!&ԯ@YbD1z/y$^4 hUn;?QhV/_VF͚5 w-ctkT^ܭu^=yM"]75qӳB(?sޡ햍^/tu6^:a6ݲL;ژnw$].r-t>Ry[ɪ=O6ʽA.ʰLl65aZ&cy1U|iwfanItݠ+g_x6{bHJs4A+P:1:E$3AR:kmy71@ݮN+B  ^~d؟}w( Ƽ̸KsP21ؖ2A1C\fj9=Uze|Jiv>~[iykccin}@u#ݸ 5(ncΓkR E_?Q`mE "mp&1o${>w|m#fY6!m\h;LN4abliv4ж7{%A6ΎLl֓[6qs8Nq^XvRʗʙviU5샺/xny^~>0ݺP} yz>I1?LZö+?Pʫ>E@%%+OpթJ.siQmAxx,H=.EQStHY]^ם|-,x*:"{]K}q.\W~y=8?޸ߙOz9^}|^;Vx+^#Y~8:\(I1~^rhAwa(^o^߿=jCUCΫ2S6x{7sx?i Q*xsBs3'r)4zme zJu< vԅb{Ы~W [34=:>X:*|z[hL<1T wD15~JVn{Z:pF(e [uZms2J=Tp4 hKgoYIE09˴sn:q2ڱkd\*+'Վ8W=Nٔ$1M腴*rךl U9w!Zeuh#S4_lg6*τ_iMWUbhep7]xvO(vk ΈhMEZ_"1ҒV`ȜHHb jޒ%t\9y&K &!CtR+e(,K[ tΈ@h)SRXk_nqPzh|s=6WIZ!>WY‚t"`A=ͅm(XC 7M!0%JY ЮXI2Jlʆ5/&\/u̘/Y!fL̘ZTC>r^;QR&!@)eV. Cٻ6neWrp6 rӞMP4- 60jDTIvwz{mG^'rA ۻ,p3Γ!BBKIf6H(>刉b:s"e,8+It/TĹc덐KC ױ8QBT"[AROƔ fFFpi0d<%vfB`VyH% tE1ȕ,@бv&튚gfr%r 1y&[ڜA"lJdcTo$Gr$+:@҄51D&("Qtedꤝ;vjs,vM{;Q4ݡD.|!QMdjR瓓YH-ӷ&.LMSyw!%\k?&-bj,@FYS*2/$tg2z:NaEMߢs4S:Wp?7Es7^Њr@mƠ6fcQk44"9R@ L;!c3h\&Uֻ t$$m̙& B\j 똙*rk!xYl( KlZH"}HPQ4GD䢌5MH*VE")=(NV` PJ"BD94;wv- ӾfJ1b͗:gP}?KSx O[~Jh^'TA#7&ߦjǹRK[xZ-ʟ Gqo*:c3/>y{ܐ!P2t?ڛC8=7c?:ۛ/֫sFFUsKޥBjZU73a8h\æ=ͦXʊSkSoY5u 4W4 j҇]_{NWQ_}4gs76K=?osx<\pƴ*ͭնgEy΋=u4,i{2 'TMrU`cjt>M7Tјeyxy  |ߖ~Ͼ}}9x^xF3?)m"\АmE~{p{MS{6S:..ݫy1; )Q 7˓A4DK;zS;֐W7.:v|C{xSNeK}VY1΁*{\y. (<ȳ:֯tS;[0qmggvegwV ʭ; Ь#,)ko5 CWLĜnHK)2#&c7[2,)Zv5kp Ea</%0}&ƅ fHQf^vHfi3Ϡkοk x3w~w \8tcao.iE>Zх'Rٴ_&;ץ&1D1JT]Ho=:U>c(Vw<0,6%:8{I߈Z&B QJuLUuSe}f;W9D{2ɶgZiqkMLkr:/4몸+]OGnj6jCWpe EVCJD!aQ2e;%BDᨨm+-h!Hp%7/o ' (!3$u{2]cI'xoO?]:AC0`dY>sHtz6H!"(\3I4E,w5fP!F'|A#,L)I@(pR~dC& c%LەxúV+ ļ2)󰭞nqO Zu5.,z7]?*ع慔Q GS+Zj.X+"KR}Y㯵1۹gCb'ScTBUT\ 5F12YZ[eE'r6a HtKMYe\e4AI"ہ{n1! (IՔf&s;%/!.U1GI\-ͩ: ){Eiv3a\%J>z(& X+8(y gdHEbl4O,7K>yEG^L]1C/dvv/~5!~c)Z}2KWOؿ?~Ү%'9d5^4)$Yi)7z̙5`sE@Al"si˄;e'o :w@S^ +;AjyfݿZ|:cEV˴l<_ZNyQ2=Wt*{lr2ۈ:MN%׊\ۿz?jn3ww<8|+Em Ȣ.GJ 6Jmcs1GD^CG=MVd.䝶&Iy4E!JKN$m2+RLJ2OZ?FC֜t-&Lz{vx4S֟׉B.!9%Ne٠%1-zDxD(cy5{`3K3@,Hcb @dɓ&v`DQ>x~v;-!4t5Ήɇ];芛oi֫[4'$s!8UP\e-$ICBDeZUJ2O73-b=#ϗZ]TހfL;f$lR>hsRph1ww;ggQԢEg?  &G-]@G_ċ.z :Q'mZ|0\ CM2]Fkx^0-LPJ+!K6u IF+@U&ږ f(S.x L :A=BC$#wk7_ k$[{~~j_Z56q tD-tTGn8sꙃt^͞rTv08 dv@t2i6O%),%o&,&a2*-No0stitt8$(g{=ܣq1`a7Owsz^fq버dˠQ0a<OկGi+?ܢ:89hA{PFmymR5ros*5m$~/Aɬ~KK'%du-~o+­޵>/IiU|޽x+&j=2cTʻ%L -؆K1Ao{V-AfBr Wu v/}7sUyztn~?'~v2 un_}[w|6_4݉(ۙ/֊׌? n Sδ1AԂ!w Ixާ=QFӒL9#gt8پO\kM`Ynm4ُX]!9~ ryifs1Q˖·vLݰ?t0Ln?m#D!{Q|7zyjR#Ag6Nn5(OVW߬]<tuUWqWlwOڋif6T<@b-rvb{j+u>0y{&I}lf5,?7_S_L'7)?ӠZv$zHzjw˦=~tue$\j?. n{}0q4-)|.hizh)Y33%<eWOWT6SP1*JZg<1F':#DP"(ֆX).X<# ЇX4`[Bg2q^NA-;]ON*r1e ꅃǪ!_;t 1UYjQ]uQn$e} ϟ<];e#i}Z Xj zg{1/+[ co1YKhqíD(*z%x3|V[BCyɅAɛX.:2mME <\fc"^H3Iؿ?gI F^YKP)J`LhFсpT䢌 k<(HeiT#@^ރT^pbhx(%^yrzC}PWn_`R>8 5]x̫͹ /[Yd:RBCpp+OXpC0G߿޺i7ޏs+&>.xSuagiremq@ !#srp甇4pO/ں?i'R@x(^ j[U]x6 ǭ?rzfl7~b(Nܵ ^Ѿ˿(ttt8;?\?T[Ygאp,U|6=\Nts5l:*#w|mn{V&;a)#y`yȷh*!=լk[7eTqSA\3ĺ*zaqpJR0Hp:=1BIQha9+}K|v:  _S |.xAo${3x!P{@ l4VA$)_, Wؘ (U$F_{YT<,lzUe$Q@ }Ryt`""1h4ve]WIoe?ڧrgj׾9[!YP% K0߁PBQfI Hmonv'rZK܏;m' C "P;BQd݀G>`܀;r.8ZW$0%8<ʘ ÓֈV3\/(X41k+d0*' s#5V؀vpRP AZ%EMi#9EԿ\˪FneV`C, CBtW6C[{\ຸcnK9x~?_V!=g%~4VƱ^ּx;ywQιG?n_ZDM(sgwg6MDI6=pOװ KB]9ZHxFBJ= :)ߢ"Qxo} 6|4վ%W oZ6k7跣}D ʽ7(M[T;T^L^I|NJ#%F:ZKo$A0# fdk zGx8PBKk pZ)CPYB8K{j%L:Q+ZBǷ?J\osbEnןB)4(pꗷ0x]^25!B=In?{өww/->L{?渏ka~S>&SAVΗ驷Xzޱ=^?mVQ{"UkҕqkUD<2={?Ac~>ۃ] ϟu~#\"TLl6|f(~D8k.8rwwuݳ?h"<ՄW˗E'kxxn^_ۭmx$K%]-hУsU?tO?oתQr,_l!2~fq16:EWJ{^{3"L'ٚ;rcu9H>I.+G]Ý`Jӷ,IIm!ZXje5B-s6nW-H ͸MO48(Q'oz.Z'@R P %Iˋnhh 1P Eyo@Y"1Z(%,wV"鞕,FC"!u݋ugϧYcB!ϗ)a"*H xL##wi3YJO\Q(* W % YT@xɍd&b/pͨzV4k?6pr(dLjuDD$5{_A#x}۝YYܑ'q܆7oٟN5K޲z҄(*?AFfQ M n*{h`Р&(-N1VAS#uLֳ k 0)(%X +/b!/|½bኢlo N'4MX6'v_N' Gl"r<.M^Z nBʊW D)8`8 s=,&G I D̢E‰ ª ˆ]Ϧq j]Q[Fm1YṏS#b%sHiB[$ PTڅ\DT!hE& 3A[ ^L2F5h a1r֨_f` "]QFD9"∈w&IR QQ6X.ՊᚯA :g$QJW-% 6)MyԂ3>`(PTs0!%̈́NSr`Vcn6v6FZ4I`FdTeSd2|/22"r8n= _]4_k]p+ђs |4DcY_omTDlT3RJ[y[3^ռfQ^Ff"!PGZY63 QqEzhQHϟ\+S X[-HN*MZ2WN@vd{+'{͂(xXBxs!/̉ripCŊGj4A ED %LAfH(1\2VqTj˫!a6&ΎWN2x<%]}{! Y_˛ճ8Aqr AKHQ̲ 86HE`e6)RA1MӚsprMG5kL+Y 5ng޴d]9$z^21KaK0˥LWY6K~gxRcbS nJ=-s|̑#B'`A<"'iPmfO= 0\ cEMiUOKgX SVYrnޅwls!764.ґ+@hsե{3zbO?KW/ɤ;{-lf2;7 $XW{._K/t\ä/ܹIx+8q]T)*P-+Ię1>W7Wӌf>Q&{H'+݇Ϟ[ueħפgn겎Ua~֮/O;o'Vo_0$.~feM ~A_~LǗ+-dT{,/Lzyةٿ_k/W"j+ |~}MMc&,Ab1<T"+.Z mlpF6E8 b`fE3,$^Ƭ=g1yDP8e%)+]hJnO9L:4ԅ,oֱҾΦ0^>&ợU7fV `&ALKOC dK3 "hR_gyԋQ)n:?,jPy# 3'E#.0TF "\?Ze|׭I+r^?f9^BQHg&C_C.ZJaqHc#5ըMӆԹ[u:b,;QQB~+Y1lO=BQyX]gt7o5Zd2c`M;8h 71{,(˜q?#A&%^z.KR5u2^pӎ,Hӑ%p;`T]̹C]̹CVs?w(Roใŵ u㒺uUeRUsWWEJ!Zujԕq{ =.4z^|2w RWE`.F]quSr1R`ތ u BWE\/E]i8wuUl՛TWB1N"2_Bie箮 UWoG]Pe;U[Wnu GeZJC}[C@''՘,{` ͻX_?^n7ۻM̑;wW?÷ Q\AO4ze'W"VE$ H!e]rC,&& !_S;{wz"͝yq|e~[=r0te= sp:˧n?(u O?ju9iI?( 'Ttx+t#1سi|ڴamڰi6O6j(²-nUMWw^%A^!vXً1q3k¹YQJI[34&dɣߚA`V ,?53W@ [-HN*Mq5× C8+{ XKCDmF)`! ͹愵9DHQ|mŸW`?*!CdPU5!z+ճcyp?% AH3-}f"E36 i*qH:E4dBQD d\B\Iqx=Cm١ MumLx_F%f^b.cS"^ m {X)E/7< I˜2[ZSۉLmQCY74BIv2db1Jq+`d҂pg/ف9\۲ٙ 3JWrkRfgȚQ:JJ%u\)Mnؽ؉!j3IgrCT)' "|( nM~%g%:h"&RKS\"r)NKI 50RDyo@b<L䈒 (Ja٘8;2ljn~ 7>Vncл#熅 $Tܱ?ښMKwpT`M8}"Ex7ܡmvK.{{mmvrh~O"6W]j7(\0,MtRLJo7Rf&ӽ{K,Iٺ{I+åPB󡐿Mn6ۼ˜G}/-Ϻ-en-zŕVbmjN?-N?/sLnl wh/0k.)ōT6`fEܣAPZF3+Rr3{̤ r]kZw$8m!6aS4@ĴNhSp).Z mlpF6E8 b`fE3,$^Ƭ=g{'AZS3gJmvFfL0`3* 5ʁSVrZM1qv?dOǟSОec:pBdՃ?|wԲ&X5̪"ieO{ HEub'ơ/ }$ 4ϒoj8NXŵ:b:Z!3)3|BrB"P պF֡&KNd`hD4`t#x^ⴞeB@LUEcwXJ8z(%#b 6(m#˿P}rmyeOě͇DSX"9<)} DP$ZCIh~EjDmQz2QRrwE:іKuFj8DhjJѦsb G)EȂMTYJ,7PbFW=}Ybvfn1 a..ǟ^#b2T_6%C$:03 xQ1 eQ(Mt˖##$\иe -cx@<KmTM'RPiq(⊂8ODrӷ~`촥Ǡɻ-bqVzRy5SOG猴"2Ǩ&ɓA{ ((u Z ̜>8%oO3g0\Բ|k(6Q H 2N >(8 S^i-i) U[%[8yd)fxaIg*uRۨ S(+i1_8[!Uբf?P2tP;)EM[v={?$lH_j'n*|lwyթ: _? ~6^䈖]iK {1yoBw,z]CF ІnOɏ nRI6){%e#Nnpf~&'nǀ.t{ʏ07\:]M5}9~|ipןS&BJ !!2BOGGdigϟgyM83q9V9 U7{9%گtpBfjF^W3/[/r4% Eʮ`SfrEZU?d~>3)P.ozWO3Wo ђ؈rw8M`O3W0fYFpb,.Wʱ!dR Z$LhFradzdmFLY?R4DuyUMg $S[͜,ph`$>z+O AwC?3?D9 >D:v/ fItUr_Rnw e'5ylhqP # {[{ﴟL-B]ΫYl4^hlhxFox nLWSsY/rUfQP8S$̛tbC0d;wvi [r&¸!]r,.;BKKarAV+F$NfcTr0 l\SızBDPm zh*PQ/G9Ww }9Qϲ`W蠰HD,Kvc@¨fyv5{@$p6)[@$h8; j K1dTe'"+DgHrRPے\|r}8\~n*xi7#hV#6zw%+3#Tx`J5 U-J툷zbb_Mzen,9{θӬ}hݺuW+7bd٦NS]xk,RJFZPlcNQJ4g+motu|da #m(x>ɽ-CwL ̌)eY,A8u. E2trk/KE^6+Wu9Nj-)Ѡ^x r :#f ?lڤж�L~8 03/C5NV˶>5`5' r-ύ'L aހ9zqG׋)Ƽ,fCP!N՘ E LRD (VPId(DbY-FíƅeT (40(9נrZ}Sa_ o@ԆEfq0'h,ENIuc#>k|RJ'&,4M7{=䇐U~; (vCHwKE,+3B=`3NmY+0C0C!\HA)cpmn: 0Ny56 VE-6`M2hPPluٿ(zi||?N>lh &RQl_ ..VrNTZ]0JVʫN'xj)ȉKat+S0ySuwVޞ\kgy] N. ˊ` 6wqY-Ջ]rdA#_*ELnR?µ,#ې? ULR|п-x1g7rJJYg/Y7kݻ2bIɩf4Rs0rq1Klwp6(PR=H LЩaBׇ 8yfSSŶZ&ߺZmMrͼwGż1ss[bƟ/eoRp,8H- Y6 drg_̯&F;6r(jrHRIm0|Hm1>" nXk y5\͑ )>e4e ;'Uko-3(, I4Nр) lIk 5vW9&Jg8NDpiiHI_Y `a @%Di>[vU) p @N>)K"2BP.!?+}{fva)R?`t05?8o/{Qx󯸾a2yXhg v]T辘% /+=*eϺ'?*\( O$_67: n]u[N<_calϥZoUW7{S^*ldT>[r5g6j6 Ԅ`:Kۅ 7?feiE`{m1:/|sP^ 4zeW zӋy늓N-\tasə.7J܃K0!i/7EyS`NN-0ƙS}bP@qȁ` Sn,|ʋЊE#QT8tm;m8\:2VH `dQyb*2JyaGkcܪ$Q1`RgG/6**b)7{ k帓` HcJ mB䑞V1Dbyʊ IB%"R>Bz0-X%JZE*NPLcʹ; $XH85PRq*!H@b_#_* }R{c [n A-l9OX E q$Lsk7dm#g4`+$T΂"߂-outo/O>=97ۯT&ŒoxUF#-uD4CȢdd;Eu ͈) )bmFO ɸ 6u3E_v}1z3o)¹D+% Q)KP "E 結QʀzJ7%F޾zjiд|F]3ZfR|~7d`I:}1\eS*/exFe,:u~?&'c`bҹVpb#V:0t>T}E}Z4f/-/S:d`k.PEϘ8PA @1N)aJҸ{:WꊗS9.Ck| .oCj;D٧&p_@&P@;YmXwo;5붾ܛ~O5U]ۭ[ =n53pWsVLI\:րfZνs^򅗨rW-۲noUGR7g#7T-km#GE.wqmŧ|dw;G1"K^zX%[;VF"V:ED2"{Dw 9xvICFhAs0%|,,,6F% dgB JԨ1%-o%eDlMϐu \ru(u˸{\q <~`$OL2gN ` "#$ei|YYH0.v[i~x lypz7 ;"%rFAG?>SQ%&Zfq%?EZͻHid&B\=5]n狡\- ެsͫ&䀩' gۯo6NfS&G8pq ِG\9\iuߗᨹQ A)LeLS|j|LƸ-d}9l:u.jeXFW%yP}X굸ɧqu}/ˤMȗyG|N%*k!WYS1KH*3uɰcqDͽz:If=F'pj`-jKFh:a9Lуt &̽Z|-5 i!׊w}3ُ}s'tv$``[_-/e b]_ug e6yKOz_6@ҏTRWN KdtNOBR"I淗,/"vBDzL*`&%3)Ǡbe{gTWT1K $/" MzN{ܬtIΣ,@% s9:YPDJY`*8/4Yp3MX8KF[6DdڣA(&)e)reNH4x,izRԠ Sc[S}qʺ:;2VL!̙O6JɌI):4)tCg}՟Y˒UwL%zIR$!0 i %6"H{m &X9=>tt$A,3$G3;EM._sIY5=Q/X8{0?l?/#?4->xN6#9\se8Σᇳy߂5`<t9k 뼡_@<6ocKRmn|[3b{3omf}#M84D=Xzju쾛5mlU[]Vݫ\wi]vRGJÊ?&/}Y`6L[EW\TӘm~x~xg4Bpt?-;~NoьRjkbAyk[7]2pmU}­{nG&=lG1f6sf[:T@- XǜCAIJR*&Apǔ3cRZiE]">"3`I|ҞȇqA:jeGԊ8o&Ud4W /)@>HH2FJ B +XJIm - ӎkfUd'lAR{Aow)H4奻ϲJNBdFL 5ѿ \@05DuD Ovm`]UV1yelܒ'D8' ֐J4KZRW] y鳑.cpB%26(E0Lȸp;CEg"H+;:H/qssuV/ꅛtz@K1GܤJ&+SձJd NsW-c2YKjѣJ^[8@0$p>KIq#-=ybWhs(%+dЬCagM"masrMd-J8uށeL`3)ZۨI$nYAZ oj]VeW$H2GQ&D,4j,"h)eW,w%l^ K`WI΢qSO18MjX RdUOZpV ^@l:JrkkazµPsϴ zyŶsOO?䡲LJ*UHf:P?ON7@uiGώNȝk׼x{yiIι'٦KKߔO*[̞tϸ d,bHãfº,brndS3K< V!.D.>MJ(/>9MkvpT1mqoe~Cn7t`y^*w]ΓxҽScjP-bՍOn! ޢֹ3|O}WU>ͩUq<\ *Zv"FY0J:+T\p$d[} Z۹C?\KAOs挀\&~<\J=b,0ђ UAM2,wP]n#Krbf\" a` Z( h"폔9ZZ"ʘ=)`z]&gp]&%c\F١UX5R/ݲ Ee, k]ff{fܞ>1 t͟@p86^9bS(*y?BpΡ䥥ш$ (.BVRxSR*bfP=9s l dH E};3R9ۏGq<Ԯ6:EmP1M#q ^<8KN eլ :@W)Ġ.aJPz D҈51Ѡx&($ċYFЩNڹxX͜x9Ǿ(+#lqcU11)$e0Պ㚯A QJW-E}$A0f"j>H*)E2!'A VC焏#.NI?N)Ym싋2. 7t0] BHZTV!^%-p$I+". aѱ/x@m515x'\#&rF%Aײ_(2 ]⥻- 2-4#ߣ5l D!'"LJ51p2 _Uq݁tpx]Pl9T >Km̱,|B҈?omTDlT3Ʒ2:e.k+ujNV;5ƴ^i)*g$Fz͵#Pԥ57(|qS1ߌ,M\Ѝ`V ,~Z?\ҿN jAwRi$Tȁ< 9V!;P^ԯ:,1JK dר9w"9Q0 2utYK+L""B kg|L.;ʌ[3"1 8K5svx.6P9\Gw=!eX=s7/775.QPD9h3) Y kA(/.}MTЬsO[#*nYWzzNssB~s߹eo4;ѩ5NAlɼ{䷴P 8z` _ x:k\ccS;2@%8H02i1/kkkOre)r:+c3)&e}5*TRKh"#u"jGN+iu>J:<[uԭub_zzL8߾돷uvح@];z!vTQ)_B #C2\ k$Qԓ' h$VK rpב5KiJ{Uwg-C6<5KiT/DƧpլ )qC/iB`85#r"AQOЌ{sLXQ&q(8ZIkg):=VX$i$3Xq 楤乢2OmX$%BHu),/ޭT,hƘptݰ9ۣLQ蔢L{nN82fk7x?]$K/\rY_>ϸ KtF* -KRkFL(?.36Fu(Q'%):h"&e^M.oz#^vw[ݡ;!w=Y`OY^F|lכM1Yבqbޢf_<,dX﷟2 c  OTG9sOeߪȚx#R'`A<2'iPmbџL)x {)@`.sƊ)ܽa]s/[Nݯ»^;ɥM×4+֑+WS~1wh p?x4/GћOs4rc6{2Knv'# h=i7g^}μrP[9]NEwg-'}ez+W]mCsqC6/wa_l5e|sMXyFJg̬չ43CkSofV53{4ryk]\"C[J#?'nb"iu"GC`S)) $b "L4N;OylSK{-ɛ8޿0r hwQy#4Q#dA[-z9*^Ee6e`M5.X3gw5C!Cیz)].w7_e=<wx0^4"N [: a&N2:\)!XOu7GyNwUtyΥ^\gd )Y,Zv-~A)3۱kQ_bW?5<Lu&1xVn=sjܠ>.l^8Z;孝1eS- םD|< lrRWWŘAYlb,aytS}b%CLJy@qt?h-O(SG ѼKY,lz`N@w4Wr4g{t^dآx`ni_knN=vw{2<&6y3a:&}'Hݏ-NS7#j1Zf8M+gxyEڠmQYH uSctǴ`RLg ̣q '苣cm,tP]+#0 ~k];ms`{A5<@yNuE^xXuƽl9J0aJβ\~%']XUt.}.׺GK܋Xޯ+;R7=_n`R7~WcGk9D9Ē#bFXOQ*z5HrG1;80&B\qa 5ʪ3@<,9G$Q59V4q7(at󻂱a!źʤ[Gs+D]<0*?H?S5_`2E/!zaAxgrʖ^r8Ϸ`Fя.pGT?7Oq6?ab [9hŧf2.p01 b?_.,>_jφW:e?@LE c,pay:eXDχriE.uaL| j3MSqx;-eN`c$1KoId)pMۇl#z.E H0;,'Q`6DӁh:M@46 m(˹i:l6t DӁhZt Dt Z3t DӁh: yDӁh:M@4t DӁh:M@4t DӁh:M@4t DӁ#,ؑ_Cɮ 8kSBR~>DIZætMOWկ]5@ } >Cb1@ } >=]2Q>b1@ } >C♘0V3#0SN4&5eB)p|C Md2 N.[*>&%QqvS܋nU:ܽ)q\pɖPLN8i"h5&VY) ǺRk2!X4q؀u@&VyEeF=4Q=[H1FgI[^v=u˩ͭ{ipBn[#pLB3dw&I +^LM=?Q2 3--=3-8iEQ2;r7l6{ǜ5h%MBVM7h-7h-Tj64hjg;uΣ͇ ф%B1Lƃp"TO8$F툗S>=(''+EY `9Kr?gyV$iF̓=q0y#jS'ϝdg؞[W9|i{ŧ_:g.!`݈у*)5K/XLp!ZeςƑ5 S4¼+ck]}i[;/rؾia^Kځowc=}O:i>>:0t}mfIf/mO>sدB-IdNj # %Xs&Le\Ihȟ^&{$]̢.&~Y~m¢2'S9&}ݛ1W|7>ӝ.? t oyx,r܁VӶDPw_ZI[k捒yrR8Ii/CcM!m{lz'/Z5Z@X+"|Mi= +UrIr`q qzuJ.%M0 )E"cp[,F3']΂HvJd5ry_O9ۨ7U(B\ 'cJ@lJaɒy :JVTB`Vy% (c+/c @PQs7wx4+UR]|~ٗjdS'ٸKn&Ӛ\y.~fY :i*j-[~sVcoIZJDYY"A"]L9lts 5h4Zth7$.)k#Yn'nKf!SLJ۬4Q"g|`RIq Ʉ@@ Ns %DF=OjOrq '_gU\T oKwFhh!ƩȬ6K}ڕ%DBY4fub_aձ<ԕO>}c#>03neU IpCe?*%؇T8E0P5V X#-3'X(2ϣ 0E7Q8&2FS&&Ȧhds8]*S䥈'V(HӡH)xg+юr~)*df]PNhȼLFOuI@12h;xT1 @&WwCmE38m, *Q%䀥 02X9#$pهwWqˏ˪ó.MӲ\PQu7gh,6X'M`2g.<@1 stKH?[mRo2E<3D܅BB$͔9:0 elFH{9 (T{P^JK@$tEh rz#i8QfWHyoע0;kƻXe]Hu4kZ񳔭ߓ;RpHn5?t%_wreѯ3:qHo}8Dx:Z#1ĥ7iμ| #pC@q'{ C8py:˃AfunQ,9Uƒwlăڤr m[qYAi(sng.8r4CީrtZ!^ѾVkpPCPiЇ *r}OX(U76K]8)%+UM~;LGˋk~m]T,ϼnO|w8Yx vũ8rn^sd4=xh翢巙I5T#=i0{;^,oicfbgzWcONjΘQ}Cvy5%-B<򨈍C~GgK䦂[nhMcytyz'DW߿|w?9Wo^GO4J.8d_~y|2__xC{ mS&>ꚏ1ո1;2яeu|x@˕$x\W8(nRTYĆp[boH7>h/@xl'D>c5V:,Wa#1㬏I$&DҎȣPbP9gO oβFΆkZoh"o۟cТ*+9P%}+e'y!ZdžN<-.TmM2|UCӤnqv,ꆇ+%ȘGΚ-4kEfdf+RtYFҒ ~/I|4o]9 x2^JVc,8Bb'r fzc xmmA;;1vw`1]UwSqo.شޥE1ZQ^aloYՐ|U-/ͷ}U%c8 ˦ f :N^('B lTs xfUuUu,cUZ dֻQ[ZgO퀴:VADg)AZ+e2r--/dbyC!v%mP|bE-|yo.2MŘNf"XstF<ҕbF./8?Q+yrpaT- -/ZBj*lΗnNj%z8-,EON:< .]]zu_+t?8Ҿ-ZٰvWu4iyV7?Z-L`gG˺5l:_ #`꧿/Gވؔ1:<j]]6> &T7 >LY7^9󉛦FZ)dyi0 eͲAAxIYqraP;wEmd^: FpR*bɁi wPJ)Bd*R]mVC$ &(3&"\n_Q̅U‘`q/3ef+ӫsπ̻Аx٣AtȌ,q_ו</WԸP9 YRy^ kL:M:u pT5cL)3,@(sY$hȹ7k_F̾\c S>㭋N? hϚu&!6K^2ڞFG jADMh^Hd*kteZ@;/nhk/OǵJX8&[0I%̋&֡o'%.ΚA*c&, a!cBkQ * FsȹݔB}j4gG9w߮i4'Ό? 6` v+<LnSNH ̥+tx7rBcMufd%X =3k$nytM}#X(JEXEșnJ&_^/#"3"Bqus,KXt YEN*AW$ğRLbq4C4+18V+ hor' 6#Rкڄ`, Gm0Rd9V s"LFWnW/` fWMVyl]ey>->.42v/|,wf\ rEk:p&Lfei$`\h6w.So />a]u\eZȴ$!Ę@Dיm;o^:,]N=Ϣ]_6oJ]XڭWLeOtoT_fx{J_u;E*dQ@a4B&yBHRғ'q4k$:\&%)"3aJQIdQd"MI.-2d(A堕[E( Kp¥NzNRyKBB)R2'$$ujl c38 5U>W5vf}rӦdyz 7 vw\Tm`A֟Y2*P\:Y~z=nw>ߛUT2j6}R}W\+|`ow}O7ܷɜ'MY{{*^.֢#,>j3!-6SsuѦ3v܈[d2ݳIw'EwMyOSsݼWewd!usޅ!"/,E w4 4&%HJ:\I)8 %gTd2XkS$=Su)I?WL{K[*ӦTIeK2Ke!\g Ά|c@p&`0^(8KrJzgp++^S79֞G.j?/+ΗR.!}}bnSSy1 Rb&yyU TA/*?,."٠ 2Q$-\HR.0I/ՐՐՐ') +W2OK,$TBYk!`3rJkU٩vo=i =A? X׶{WsJ0Ԯ< /<? rY[P^2`Fd ]鄘41j{GDWXDCW WX:]1ʺbU.""D4tp15tb'!]iZL'Xu&"(8Qjte@)w9SuFnnyE`y@Zd7`v Ks]7?\_ͣ[7޲I]ȼsPI* 5In_9.'+ Iw_]]iKCBRiOx \nb%2 ;imiNji^ZȡF vZ5@rͫiGy%)G X{KQ6`dIQH\:w' h-^B=˞qɊ]MvqP=[6(swL`]5d;]jE؁QBǢLb^}i]ۆ YiqJlQ ngyo=a5vIǡms.N5@j-k)MSFaB[{#v:hoӼ5 ۜZX|a::mG ہ5mW9Khgw 7A~w?}1+bnSz9Y$Y8˞_ُ첡Vخz*Ag瑱ӆN4\ziFmQ/ma%Uܘ:Pgj|~z~_B(MhGh( v ~~#Kǒc_L{%81CQʴu>TBޕjY0nl>MhqZy 6#Auͧjp.ݾ<"NQjt^iIbFCW X:;tbMt5BR^]` RD Zp0t"(DW#+N*]`HW ovh:]1J1ҕ]LtEOq]S숮WjteGWVt6""Vs2pBW^ %]b:dk׹X*1AF)'ept^l+s/jgeKr{Iv+^YvIs}dO5H/Ϝ5L{r :5ŵqj%; e&BZlLH/Y 8l#{nFQ~ރw/InNNiEUphavrFԤɽM= I:|kwVpOPZؓ JC3FM[0"&;96\w>vhQNz4zs&8DߗۃZ)y;f`aL 2]:P{]`cT4tp0tbNt5zhQ鍝mNirٲٗKn11dYjZe 7W;~7s/+Sm&ԲߪR<˒ք¿ۧꤳ }E>-./oeq߈넸oGj<\&w9͕r|i͠7muDW*܆bm];^9疵n/?}mdAWen^tq7{n6Wsv݆ggq_+زnVRfr@}DDӪ G|<ܻpYC7٧y\=`W:G^)|1\sB$/>nx&Ϳ9O}]鲠hW8-sHI=%BҸ M&B9o^⻻ gh~{v^|l^Â/[P3hnu^|]` 2 WpL~Sj)R{(s6PzHCJ>PL)r چ4! K+rC]Moɕ+ ]%$0؃wî6!D9xM=SZ,]O:H̗"20w: V>_WܰP*aMҩXa&U4]oǜ̢u)je KsPU,QfɖV1I;׵aDCt%'/ h1Ѧdݡ\\BQV 4Zkf%;v+5RšDiY{n j)-x,fL0&hT3$5$-Ռ buUF-9@hbvəGv+`o3m*F)]&{mAXTieѥ LFa!?BvL4Vi4(; 8QD) A OD>_x(a$cbu±e6P6Zb!%lK+{gusș0]l{ W@*ﲝ<)-[ݰQ}XC *TT@5zjݻޠ\pT6#Z}A)Ғ` U\ *U;0V^;*9" *lBGkls:- d>QA"Z}W:a j *6 F򌳛f2 VԠ kz z3|W4@6[A0zIB€(2}!iNFpW.1a |k)@x`f#aaQ1uKs:T՚I0μp j[(e;@gi!\ :k¦b8W -ZA ][td"wB=A/:k:uEs HY֢q܋. R^[ȼyt0Wz,IˁFkuV"z,5K͐ք(ѱo'6l=>TVcGbE |f ePQ,KT3f%4' ǎ1KTei58!a%M@|7zqsW֜sO^^ ZŪ'oG.0mFL`-$L>g@uPyPķNl;@6>+M$]͡-6v 2r i<1 - .Bp ) >@(&=9t^S`ȺXÝ.̓u9Z|OyLP( u6B[!q:Cpo},\-d!:T?QFjΫ Pnzu(iyT~zϯ.u'kWS,UǑ+/b Vb=ѠeTAԔKǠ-A^,s` Ѩ%@_Jf;Di0eP0F/%|ru֬kdž@A5DG7HX}Vv:m1E5h3VF/Q7tFX&>) c4Ơt% VUjʐ pyPZGxw؛G7 7?EeXEcA9QgũB`vf]S@3`5!]td15MVѠ8Yxam@57KoMu^Ipd,XH)40|p_lkoynms(` `C3ݛ-@ 2sn`g=gNA{H7?9hF֞ek0sΪ6YQa6fѲF&h4vfc/"cw[/*3&5% <%:`Oa@rXr* ~Bg؍[\{h.Jtr `,.0LM]ڃ DGGPzl=q >ߞbDvR'w ìsXѣ/z7RE7L.Ba ۢ Һҝb30|ףT],P~ xm,mT19cR=iwml)X[rrJ@zxETAZQkԦ5N:;1jj6Ԭd#F_Tm >@|7A>b1=i1`EXk*; ZvZ #+Bj@S Az|K!/ `l'aPJ↙ttk oޢp \8 чEk6\WKcbX`XKEw,@jQKp$b%˔iz a?l9sG \5G 6M'ʵCߞ>ݡ{_e}mv]_zsu}Sk[vW7}7xB>z^ۏ㩺GϮ0._W~OCFQvf[ѐzg5_\vns[9?ɂ[>/u=nz~Oqnm_}M{;WfVNoS:lW;d‹plw!w9q~4Itf|FtkcV: X܊oĭU:+ qS[OFVĭ$n%q+[IJVĭ$n%q+[IJVĭ$n%q+[IJVĭ$n%q+[IJVĭ$n%q+[IJVĭ$n%q+[Ir0k[_ӭ7ĭǭ2z[b*&Vĭ$n%q+[IJVĭ$n%q+[IJVĭ$n%q+[IJVĭ$n%q+[IJVĭ$n%q+[IJVĭ$n%q+[I[ƭ[S iEq+o`W>n,) ƭ&E[IJVĭ$n%q+[IJVĭ$n%q+[IJVĭ$n%q+[IJVĭ$n%q+[IJVĭ$n%q+[IJVĭ$n%qo2FYWb+[1Zܱǭ:ŸNĭ$n%q+[IJVĭ$n%q+[IJVĭ$n%q+[IJVĭ$n%q+[IJVĭ$n%q+[IJVĭ$n%q+[IJV ƭ~_pJ?2޼z{;z8pZpܵv}nT.g$mlpެ)ѯ'O?=hai}: (Mt)iAMj-tNc+Fi ҕ&"jjbeG"]yuN+kWsF J*ɮub!;]1$w6:E"ؼS h5tpj h:zu(:AVQzۣۙe`}+c#ٳ]_= m?LͿ)ټnsjh^`r^'6ڜysk8/徹Ηlޯ[~/o.kUo2&毛%(~WE\|2(r|%hλxYW;\;_]zTTwot|k~߽5֌~rQPms_8(Z ]1\cBWѱ Vʓ :p\]1ܗAhIc+FiNiM AZђ9vڟ3xtEĻWDW xEz62-+FiI*ZqKdF3ό~3: K9%b -p*Yyn{up"pB"G &s9k \%kٳdB\Ez-#؄eWL\H4[-;ܠ'\u@U[ϓڿAzԦ<7$sEk2ZyJɜ^_ $,/9`}ꈢ=A|":Jrf[JwpWRMU2Usd%S\A"ZW`vո-pS-/9J[+PEpKZW\*WZr+uWoՃZ'ߓ0suk a~p$7W+[W`0#5p̥d-?{JV]EILZW`J*);\%+ •$r"j=[WZrd%[+E8ѲEpd.md-;{JV.Mżw-/OKk-iC͕m9Or \9ϓ{) 4mog25 ܂qorAD_<-7o)24\>ak2ga7(!/=Vl"t}Gy7.i? s (U﷍ i K Pslz>Π\Mkr;?癅jV |Ii POf2/Gnݳ A;HszcW[cR;ĬW逰  |\0 ]bx0-귡st7ab MoovG*~kӢ[ߢ%(Zqly6wzl9ަ f彙\crUܼ'}#3~g`Z+?E&wzqwi>T>џ沋Aj/M+7 u~@ y` 9~BAz#NC3A"qތF{=ͯ5ʖŽ -r~62H`BDN3G9 JBd~jVS)s6ULHBQG8e}0z -`1<)c"ҵ$Q?k'_6Э Xc[Z]#[%,.R$ʕ&K/{!z3J`;Y߾;Mowvbeb0!CsF\'6g7[3bt=]Uߍ(1  k2nd Dd_#Nml#/]ե_;8vDHqT |ڠg}::)~< Lt;&Y!F? Z!Q3vDJ֘Gs8dzئ@IC~d"#Jьd3k,PO+Lw9^&R^&^&b^&) oh, Dce\[Α%2X"·2Qq8M(rnՑaY\ 0kZ ac29yӼLEZl -s(P/б0Y4HJ< bF8Pk)GH&d0DѠD}QE>Ǝ%$[g+N !t>`gI\*%X F`<3y/2]+-ğ4 "DǓ}ib=7TC3gnvS+Vy?]sr!DGOVZ5FJtC)i# RbDY Cj;y-Xhm 4wFm6IP1,%! jSE'oYOFY #Xkmzq5Y)rQEWtDk_cʲ8'+,Υ ,>A )7q^9ub d :ވv;{T(NHoюH7~2fGd8k;-T~iȕkZ}`_.ę7SdIcƢiؐ)U3OH>Ev>Erv>EvF>Ev>E)E-riY{8ZC$k1)Rbu4D{S(sVA 0PƽuH ÌМg3gU.?>͟8uj<κ'֤u>yKg;v UMUS3N^H^SxpZT*)'|N &cD3$(j8M҆jA$22 TK)5X:#zbs4ZAŞ! uKYZ + Nj4K.AM;Ьr|tD0M:!Ga;)ӁdJјJ+AHtgUy~~Q=\c6ήH6riHMR)1Y $HY*A]1ݾ}[t S^i?;iP@h׶$#<1rCcuY5^XE%B ř*Dm6jP9"8Rx:< l^X_:ֻN_YWEMݗ4AeAie׷ /ȆR :pqp9x:O bDa* h"*D‘!9 0^w͋St6ιNYBr0Rˣ-?srA#|jF`s"l~?\⤚Z"[0{#/) a:yMbJ =x7xtKTIpMhrg ɲ8肐 $HH)icbs@G)ҜYU^1o;K#n$Yoxz?Ͼ{ f}RFu$ sxe0tl@VXKx҉yZY{&JCۍ6۱]]h5R8PʺA^xXlׄRZ؟ԭ>Ѻ*u̶)VFfj43QK|ﺕYܲF3vnK,xsW-jrӖPZFO?2˅1U*Ks|e'5NTM[{6iD;ZYK0X`"̧& ^GbDD=9f Y OtꊠPp3!8 (uB! P%9$}yͰbm-r?z- e$];{ˏwwԡK?zs\{EglB 4E|^dA d&r,0"z,C>p):~PǓ)žXWe@(FFAX4)reNܠDd5(#û[Tyl&y^ӨIs//JLցE e2™O6J!Ɣ0%)&N9Aa^0ۿAEKvA;4PŸd.)Ilv%D%6 +H/r坊3-(ޑ 9T9Nʣi9Uw/ =yr%Ț:=1[_~&|RGZJ8?5=J@OX6pD),Q?W8Vv!sMvǫp1:Is0qjO3R[4!c3vpl4x|N.3$Ӄ|ŵV%Ϊ]>2(A 7FKQ#MTϚ`1pF)jy\_!%⁇>o)G.A-Kuم@k5/b2̠Craٰ)\q̹ ܠBcЪWMaxL ;&)[9$,2F |9+)gk30+BEoYk JҼ:eG; -a*R3a R) 㣅̴YU$ 9i` v[钤$DAnY1k,? "qW\L}vzUe13 [L> *t- 91+Ac8/**j5Qn-W] RrBJ"e<۠0̌ *X %H"7Ѷ 2!;-A{Y{~|ߝwh[+/ΠB@݁7]h K̑T|-c: uIs |%^ =qr}с*{ֺfQh ^BB=*<6oA hϵb!i;V$0*P@3'91D i]:ybҡ S 𪂥 RRASI:2u[QEnY@ZT2:sa+Ҍ@2Wx$Pu&ٜ@ J q| /`"5Nvps DQQ9fEM<jPZQ,M)q69u% ҵ8?fh=38Ztq8Ǯ2pоcΗrxo;qVQ2e2{_?N;W[mIܦs;l_[$|RT;+ I#=䘌4ki9+e,&)C2L:QZs@yzMk^N.?*V{S*%]C4lSW3qMHqTo 7܇߽*ag?@?,U'+b?ƶ&]]o;kTy@[ ۺn~P^JճV^+>i0?Η <ȿ5ȫafD൫lwկGͷ_7k5O .Xjx:'#lz{q1?*H{{3"Gdà gcu98_A%\(?ۿz6ڧ'5vlbMkɧ,QreYrt6XrJ`>O?nCGbwۍ4!A5)(@2)gU,y(B9huCsFn`9!@HчดRrp/;N! l #C".u+k/ϿSpPH -mJGH('cZ@0x6Q]v|4IܪfɌZ kP (w鷔rq'1.RUV~ā9UjάDŊ\E˨dC 05 "v?֔DU6贵H=9hI|N4xΕImfl # kMWi/:-B\>+Qfoș_lv-Nԙ^Ndy7dh:36T%eͦK2X(>R)+Ak"1Nk*ccE0`A !iLF8ѶC[t\Eˌ;Li k6;em2kˁ υB+eUƒUs%ƫVA\P0.V(m|CD%ZȨrVtlkRfQ,lԨ8fVJ,Q-akpƨ_1v[2j#ޛ$tʜ\ FRug5_ e`bI)*#:4 %YIh!) LH!Zu@1%-o2#y2ꀼ8%_gkd[^-xq{x0CsX<1ɜ:1g, *l>"MKH4-َVQaw?kz9f|؋m<=hb7kL6a Q6vL&Xr՗`mB ڄR!X{(d+"7%ʮtj=+|#"F}+BT Qš=Е']`%{CWW˾5tE(]#])ɵ0=+]o~sZ+BD Q6\$/th뼫Pr>ҕZh#BCW +B l=+ _iWkg9>\7?ZF1h bHW˛ol~k<9~ULYz;[ u她NfWS?NaeN,q: 2biƿb~&0eiN){\?TOڷz+MVˑӱ+o H|SM?_󳖓j\?l;~W>::ˬj Ծ]'Y*)ql_uOfW B >7,GtEu"~DtCW0+}-'㵟W>ⵉRvl?@W۾z*D7G}+BTPj=ҕΚ>vzCWW}Ppb#"N(tE(A tt%Ԗ+pOBk9:]J]!])g1H ]\cКl=+Ø3}2 G"$u}+@DG ?tEhu+𚳊# GJ(:hH-6gǗD|BgZBE pYCU_c,AWhsLMYk_|Nq\cRK)gg'e+׶نٻ}k[M^1YXQ ҫ^K~ޗ|m9YYZ?{gWbHU,@`{{86V3 =odgu^%uŷa.r )qLJyᲔ ?y4M%L0fzJk+9[%TrChNN O[O|2tEiY1+@&Wzk]rn"R`ȕEVV/WJ7:GrCKvi7nrmrur學'+{T\FVhr{Ϩ|l)N$W&?R\ȕrX\)գ;SHLO#WKnR˕R]\*82+ JW_ *6QJ\ InbPq42(-̠Rm*Hv 4raRh.WJ&W(WwßX{ ^87-i7'ַ&fGi$e ]nx}sQyqlDj .Z\^\A5{1JQxrH";lN.})MA(A82w7^F ~y1̹Vsո`w6E)R1AOO1s^&ƙN8F$WJvrlrur?R`1ȕF;\)m(7|ȹT\TZ˕RΦMEHH!ǒ+ŵnRZ.WJ9U !L$WAwYJiڕs+(n&R0\)Lvq2( &W#W;[܅IOL|K|KqU}oe ۻ{,jy׸瑿ژ`Ȕ{-eM6 ly{-8l3Ur Y'W\Y*9Mh Vɝa%ؓiIacɕ "WJWєR*GCO 5<z0e>#WhٟFQ@hVP .yJqO|ez2ʵ=lG+ O%W iRS_T"`&Wg(Wz%Dr4rg+ wWJ)])04rf+=վ(׶&W"WDH'+ Ӹ+]ڕR&W"WAMn"0\).OsfPiŬ]2nkW(W";\ؚy֮ԏSXFVԮ~+QgZ@sE ڽH Ddׂ@۝&hA@ ;;S'P2\1vߠk۱y 3W0OqUUWevqFCόN7h?\-åؼOt~lorСѮ"` O#W,r.WJx3+IVpkpiJ6<\)]RJ \\(g+GGȕe]R6:Gxw-4pcthS4k+jv7&W"Wf"R?k`n])v?|}9e7л_%_^^>Owhkf%v//޽F||_)b{]R4ݡ\_8p._qwܰ|2?Lջ}:vۘ}ݝl?i͓38>3>~+ɩw؋Vo'dw_P D 2n90 8E >k<A}~x 7N?SUUx5~?^շ?3sl}_޽̎zvG{qt¾eJM`kb Q7|^Oïo~#Fh_^Cjv^orf`/^iop)7$;BK(nSRfD6d99C1JAǏHr%{R.zor3>ǎ%\m 76KCL >1v6lΣl;aa5>LC@1wɦ&`B.S[E3CAZJ)VzϽ#B]4Dzo%_Z`$KԭC3Zah6ωKAZtm)yZQ1k+l5B @ wRƌ0'Z[G,} d0E4b!o|nV+Ucl4 7 |%-~sgNol }*[=@6x6_/ &egK*(̿O !R;}ncD5 $m\ ߙ؇B,߀^7N((گ_6DZW٣+T*JKTC*AOD*I!AZ1 f 3f0f"7-[7 |1Y8rL=@zmrksi!gs>c@F5mBeNքRjH)H3͓_"$U*{uI L@KAX}myZ}$$G[ч0KKܜ=H(2!pԞ =w}*֑]Zy;Aa:Yg _F*U`#͓< KR[ aEGg;@$x-Kuv%aD)SJhڕ8XȠ-]^`[Fhcn S,&βU qk 3yFX7!7R#С˾ `TTăI1 [YsEF6[ V #ZBSct WHʝ- #" *ʄA@G6b=Ѣ"R}76 weo|#b.#>@F\G{jpB]ɳ܁6BnU{C{- bQ l-QcȄ"4ȆA#quڢobn-  ,,t$u0wLL@TR'5iC%;cԂdT 00,O8fӰF-Vz!訽Ŋ2Y8J892rAXT5,x'D)J6wCTP󼫮BA*1tj B1< 2+=d:BP-R-"#pPBywL [,]HIv1H.ECEՊ{BAFuf +aPurGwK{׋qiIb6(!9yx>3$RbZ;L'kQA_ ع]g⅋zӻB5/W sÇV11 J0$ZxT 3 fH\8t!NpMȦ9DE`jm!hiU`ư'$ aYP|QaQBE򠷒aHDNk*f0aXqu@{:E˰f#ܭHw 3xGxKNv#;QzBbޱ:&[WBNCu kM}uyo*ɼȓKq٧K6":qwpVrI| ֋^rTgHMuj ȗ245ژ n Qۚ$gT]|ƨwl 䁈:(2b UZ=KjbzFH+(ĨyBtΈ@EDe#[ qk^"8QEfe](- ˀ5qD&"/p6,bcA8ng0H!(vDMSe5)U _BzԞEwv4FBPլFi)hXSot^I"ďZvXf|`pG|V5=1ĭ7hΡ+wWoP7bwݫ7NsV&.C`Y֍;yxl|I6ԃQؗogS #bj\k1Y7jFq-Rn̘#WJ˥f0/̎hYj2&1Ax6SX|ߓT\ F @YFeQD~¡@x'Q[e! ^9 Sqp8ֹaP[ UUS ,K+Y7@R #u _7sI q4-AR ?_>Û[x3\Ra1hnH_ y:˗_^LiozY4H:;_fynj8•.Ѥ{~NCDy_sZp6}N$_iڟ4/wƽn^mnS7WIPoҴr.`Z9@s\5ʒ^e+/;x*8-`Jh%P4}@ > }@B> }@B> }@B> }@B> }@B> }@BЩ$%J$)}|@-G-)*E> }@B> }@B> }@B> }@B> }@B> T}@0mH d{|@0G{@'ʂܠ}@B> }@B> }@B> }@B> }@B> }@Bt> ̴]5>t20{@'&> }@B> }@B> }@B> }@B> }@B> }@8^'kkqzSMsL_Ry}Y[d[ʀ?_Gc[p%im)UmKh[:ҝ`6Oy}U^찭+W ձ̿#] ]*k o9z8ȝQ` @>t8# gi WMٻ3¤̺tAY'n1w9˛cI Pc.X#e ltVeԊ*Fb(\b49:X7Y/}-EeYnxY-qcO T9-B1W|ab´ ggq yU8AKZs "$N5peaEgQ[Ur{ȩ5G8[:%V_g*7bE S׀F_WEq/B(eo 'AӏuN:S-h=} H@[ 3S;䧩2.fܴFe¶EeJV(EvZPetVL~[CW.mV |2J>I⒋u@#R^L fQ::d[jp6$F[zf_p̳b7o'$]<Zt<93<XȰKf͘u-Rl)կxief/|*^2]ɻCgx8sdI灣2e.g?ܖnۤx5} F7d藤3JK_nlz! OжMWtZ}0g;]i6=eLS"Uku[2N=vI+$EtŲi-th%1NWegZ66= Z͏%NroUDkΥmUFiN$gLUek*5QWztQRtut8תMKؐ.%mP޺PJ|2xtyG '*å-t_(SM^C'~ZCW_n*uV酨|ߩ[?.Xi܏HޡVݲJ"LTTĖ^z[V,K09ۥ ƞ{;NNm)+'P!2?A_V'U.4-GNCQXtTe(nW#F-i^b$VKQcȺ 4jbbm+Rr})Uu*G7]BE^cr9 Y0Fi+٦*ۢH.6Y'phK$g&eb$geMISd!d2Z2JL-ӋgWn|G89JF!W3O.w/έki1YIqݛ^A=.C׽iu֯+$y \Nc[#(޵9=?KԻlkä&|FA_ Ǥ& ƍ4ᣱ&BOZ0WFEL2*Z驎9o<5߭vzD `S\o[Κu}rBiy0mܟ/RCkU(4(p\ή]b(n \ l0٦zl,?oݮyy֪vm2֠]y;On74/|æ#eGm;^ @^.RLP@ROo' !%zj]FF/ZaПWt<[KJ?6w:ZUo4*3ͧ'Z(F'[br`;N[,,گ59t+!hzYa ^*KtW.t1&stڵZ5T9%|iҼHꦙ%-KbM:N-RoҴJXj s\3'dWw JKhԇ{RZëAو׺;Q R we%G_ApULj  ׭ॷ;4QRjkk:C +[qj#wT`Eb:sZD:nm L}x[w(`G`(n`ZJ$g|LQ=! *XቚHT ƹ؍"pHػP7ȭGjku5^~*~yʫwsϗ?W?r,+yt44e{{q7K/;@Ђ6b;/ YDvCP:n5?ֺ,FuZB:čA[w鿍A혥W`zUa<\7o c|}&*Eee~&Jâ).qbC/rq!*g>ƾP(,Q@FO4)=2V3 YGeӅHnm嘍Ԙ`3&#PAϸ(C()+IQjvop>n|Djǩh) U[?!a!ȍ. 0O:emBv'(d dF!{[4E Uei㽼j-c7n("JPJvZ#E;J) ɸ+!IsP/h's/դt_aI ?*DEQ"c:H%5p 1i6 Ţ*U'~fx 'q,hLO9חޤgPVݮqrDut udy g㐠 ]QNǣ>DZbXZ-Ea ,NH:}) h؃٣{(&bg|3#&,^[ѻa*EIu(I-2n3\Ӛx|OZniX4nq*V[yn4ƥ=V^ G|g%Wxa1eQ?բ8EqdZ;Mѹ @E"JPM C#$"2 guQKIn]ٻ6$U;cH$~JP¡ȟARDcJLkz~]U]ȹM8 BVJHH!s)~)Nvsw>/J̫ mBxqaS5g>{Ky,5F8siMByfi6dA , R(GH!dpDԫA{p%.•P+9WI':R$paPT{.,K#0r`⼗ > 'a6!<;EM̫sCǮvsXBXC/U*X~U~e`=1Z!8k?POAd0KQ`be[,Ze`1VYE 62( FPT KIwɣ%89%x x:bG+_vt9Z]aSTF,ɲ$WW|tn6{aYfTx!V{1U_緍[7f-C0!B*C@k5f,`ZFLJ@Fs+%"f-m㚃9[$q;$pU@ErŅ"t7MnN+u:"3I*0ຜSמ)0`zF~85 0ͬ H^_Og&ȀP< lM8AqZ88Q.xBSe:E/!INͶZ`b1aTYB9UTݪf_*=!euafuYx 2+1 7LzrYyӖaKEE_H[-BB0S7(G.*y0)MIR~A?ZvVG֊M]G‡4MGi^?x6RW[p&w[5)[(Vt-&r.~8bUoç6G-^-mۚȵ) ~+L2OfMdUv_7FWϟFv ǬvnVF n?7bɛ-f^{n3oox6o,{eU >~;tg~]ndB~,B[XC\gRe8?.:i`nMFI_NrPruJ䎛X $X0lcIX%QP-QF&0aj5+&VYO bMBq}hH 6m ]BWM5kY$Ɠ^|h`E}Gb>4:T}ɠlbg*ũ 'R4Jg}oU<Ҽ7w*o_JKڃ=tI5õDZn=)=$cJ| Ud}>C15vzzTn3=(>]7b2re!tg<بUDzpܦ@7="/7lC cAI-uy$r$h* žȾ'W\c+hY9-Ns "`r[=[0bsF7EFYu{b <蜑VD$y#J s3Vk=tv{鄞y^%513`q`@(N+t:啑1Hj1#O6w=8~fxR^ckJA3U:m`LTZJRM^;j!9 m"AJQnnh7FoOd&})'nݟhR^9M\#% SI P0MdZY`8!Ӝ:{8=NN@r0-R˽*mOOn21Wi>yxM1=N̹E,Nڙ#U`&gӛð:R.|G;jd*}}UE9M.R㈔a9%*@aď_v jh]M>nVu2'1Eb~:x竒hHD|'*0ez:maj)5z.ee*O{\]}kFX͉]s)|q}_̦yw9ke-e` 0"r0yTQAh:vF|I^8;8Ӝ?T ]&30칗&mF-|8. "%|TcTU2`YIj'up5QPT"- ƈ@-M|~p J!Qwѥbmk7݀S| :q96.x `31z#!A#R"=5H!SU[d\ I XD J@TTTO', J1,,62 j6r6 j8V7%ݮ5pіOvh#rіP;o_|5YQy'J EtJL(n!MX%a'D -&GkA G @Op).}ԖU`̙R191 qƮXH3c!p0uҌ;"5/l_Uޔn6*7Phe4,gh(P.Br*E$bxd,N 7 !{D%$ؤ! fh[8"E.ٌv2K&桠v6Yff=jHc9r Přl5cWcQ"B@&haBg ,d 1+ k|@N@8F=r`TiLfpP0w슇23pC]:#7z_5j.~|G5 ؊F]?S AS)D yVh='BpFzgH Ƈ}mR-񩙨Bz7ݭ3c(2m d",J"G4(`ʃ܀(FFۙIG"\J &H8FҎY)Z)g8z `YFΖQs4=DLv_(u[uU]^n!mT7ܚ\á3rSSDB4RC8a KX aE[@jC W ۦgpKnt];v[/tnE*#RTDVDfEwZI%HD0 AP$͵=cPa(`?̖/[5JSfvpӕTEWaGiJe#\x5u˛mIERlBk$@<1t1B 9! c) XE1t;1.,Z8 DaPrנrZ}Sav,ηv-1|*E'k>jye\d0n,U;cx[q+\Oh8I2'}ipaoaq6'x:|a쫻2|J8+0CC졣+tt.#SҎnQ(\85!PpABB9m*ƳJ:~b|C_Hq[\ Dk{BGCJTGnQ䭂Mi1G߯᜸H%;E [ԌÓk Y~X}zv|~Y1[A t8R%`<p=r7woѺVsW.!(ܹ2| $xR`tOƣvŚl)չ*AWmkծgeŖZc2ui2 &q`v:ޤ1zzW8:?ޝ_xç3L?8Q K+ ?>-?l=riX]*Ҋ|f÷YW|uu+cvGn%@~}Xw/V7eqjE(P"1Z`=ymPN砪Dg02v9B3~C8jULa; p ;KyVGT} ;Y4 6~ꥧ_xS|IC޾o)?]G؊!|][o[I+B^+*n36h`_v1#%9^_%#َ\%Qtt9H~d3|yhB %R ?Qjgim'_KbLW[F!G N8wsys:zâ${&+,!ۤ,ڹ;1ȫ2NVRa*:/w8OG~#_Qˣyd%s{J_0闟Gs٫DèU5V^M.?<ۛ;O׬0[|ɖ@Y YڱBQuj;h!-V.0A ydFS)撊֊L5?1*ǍԴ$$RUM_%fhMf3jZ I}j^LO=Sbn_l|}Gh1(,\[ = =NN!r,N]QɹGY{_)TxǝqCp<A#Wy3i]-ᖼXSL# 1%4$ugQ:Rخ *Zy *x@lis?!8z[?7ߏ/?T1m]q7Ȯ)riDiZ~YWco+t"f=noCʎZ.|TAx˶;{&: $x42˄*-CQD'd |ugmM8+¥a[Ps=ȱ|9[tt 7K6QJFĈl6(ou^kC_}NN8}s2n߱Q\Y CGޯz+"_\`~o\2JXY23 YVzDb5O;dsbrm[πҿnf_ޒ-~ZC[ ]#Nrӟor6iI7/ޣ ?bGܜVk>J®Uv4-mw9 A(c1}wŝ%u@?lp E3vPJ{Gg( 3J;Cig( 3vPJ;Cig( 3vPJ;Cig( 3'SuʿN):_'~SuʿNG):_]:_{ @tN:_7MtʿNוSuʿބ):_'$1@*89$0pq4sB˜*C\$9Ɏ+\bJ;qOg( 3vPJ;Cig( G!#ttPJ;Cig( %B=np}AO0_8>_,)Ӫ\kjDb: *2\(,֡C<7רq}Ý" d ߁Hd\S*X"\lIr4Chr9m`9 ,d)9SMKS$zK";M!%fpø- ;RȚDmJGHoE)I=94[[oﲣ,pMR<7UԌܸ& (@@E(e/) q|* +j3qW|V+؄`i@P|P*"+"QIHص{<|o&zmӚ u.a-c|R 6/˥e|*?`D2GZpQ eQ$O1L&b`hQ)V֦BFL*.s0iB(TkL kUUZBhl ½•DfoҀ7LLb_始[l%(3MA9,d,> Vր2qNkbP=$$j6Ơ1("؎b"et\88[p9Nk!XfcW6v]+^HBA(ZbЈBs%v{"H )٥u(RLP E![2Zfi&+[)^**HAu67ĹACǮQ5[n::e!0k"]thH@k{8%|,njn *M-YRlpEɠ,΄Ŭ"ik[cL.t)[(!:]nlu.!a-Dȝ:qg, *lAI uH!bx(f`V Λ?ke:yo6(xjGpK5伌b(|@\~|G%5s6K|69--Sj,WOn? M1.2fN*#漵 +,hThU608k|~U7h =*\ҏEP.p0_+e-J Ҷx_)g~`w ١kw#4f腃~lnkkȸ \O'ӇvꯖzJsБ>C łL)1"M.Z$݊h`EfxbEk%''ѫ^F9|-qJ[?DzD/`: TO.Bb+ Cp:|:<9c#iAC#)NRh=K̜yլ)dX L4C >H :{IS%ᩓdIqcR} \:p! \0<+a=pDPirv`RB(xQ*d+(tB0REIᬵ FJz`Q+qO߶B%ŷҝ'Φ2f.ޑe~wteNg ϔ{0R#,w؞>wD_Э瓧mf"g `1'L*_a6Jɴd?; z< x(ȃ(jVRHY k .)cƒ-o<&2[˘A{Q /jy,W !=d/KXf)6]11={7̓JkCǙOW.]I;ikt`,E % mrmF'T 6xB;πfޕE(^#<)Ufܯq lzzhU*MSGl+zdyGd)"K5{&ןc㊑ZR $FNRdG5r9ˬH[bPXN}lLE"@'$X n E9Ǡ0@;G I*"1XfKfI'Vޭ6)GGYzӅ,a˫s.`:\O''I\zZ0y/\KyR=7*O&>l22STfH&yv9Hn3Eo()66ٻ6$ eÎcx& !v7yqU̓(i( ϴkU]⟀hfZj[̷ؽ d9"UycT^uw P#@e|iSv@_Pf&k [2-L*l 웻Pr#*6py _ǗkNk/THxfg~Rl&Yd0.> 'dSV2^"g<\ަ G_o{kWՈF@z06aj0LٺbzBxUBxx5'5Hq7:I*?7EiŔ2HWm鲏e|VP}}XǭbgLpnu;j&0#sD,*4 bc9>:85 x?#Zaix9{?I3y\^=|!VQZ#vM@J)4XAln]>cFfMWݥ9@^jmN[4ea8&@)x,iEXPOq6k 'vmS;]zg|r:|DD1Eg tJAX:Pj8Ig#,@/YI?v^"lD"^nq/zEJ%֙ H ",@!Kbs%MЁaI@St_\Cp_ sܻ|8]9vt[%;.)Bf](̂'sN{E2dg8#0"܀!(zkӘBZ3OOc/z^|im1Y9IXָ` NO%OWeHa?Qn ;Y>Z]TpMvHaIF{Ϲ"yE]qܰ3f- `>~ۨ/u!đFM(.ƱƨIS3CfV8. k#^[A3(L)!XLQI:ȡoq>\;3KbuuiMp,3+\xHD8NҴ#ZԠ>2=}qpfJRQBb} `|n8◅ń#q%FX 3 Hp9)U&n8֟)Lg Q~8#GúY>O%O ZF%KǨv&L1U]LPFZ/87W)gt6YQ砳*+K MѰr(o R#%lӼ(4wňE͝A>ۭA  %_т!a :gq7ns3b!JJ^`Ya{"{BÛGw G9`fJ6I6ѕme>_R~7](Vb-6ףNWp0l-((d;޺>v̲5 7LR~͙dJRFҫYg&ǼMx6LmҤvBE]{[ʂZDta443מyprY:AY!)͢BOљchN"b=km3LiN-_PGd\p0y`yf%̹`Q }E(0%bmuPyBw`'ee{,tg{am^ mTbn7% mO_(.Z Ԧ Ǭ 2KcdL"Qeʨ9Fb#7-\lBl[h"E !];.ꭎe촀i7Znr-[wx״|y=Bxd>Iu5s9Dc @A[ e S1GÚDv0fGtݹ+3~4j,-T|z!m_T)?F)| +>kt7a WxPAVxg߇rIšM}y'y6-naz2>ƺZK*O mߣ.+$XWH1"ZA趋+^\Bqa2NW@&rgKŮ+mWH%WzJ!qvF\!jgAvX WG\I c;$`vF\!JBjvqT.^\"qŰ2lF'j 7඄{sf|1+Nju'4X5+JD@<I h>2} K,a{ܟe쁄8[ a/\sǢ'1!M&$2 4)z뽢"@C\SθiE=yf\hڹX{X\gY:MBV2S EiULh`B,wۤ[]pQm֫lOuGѲǯ#}ܞZRɒ,)iN{5I&h$ MeTQ7Y0Ż A\&.%iaqs.I)S>jKs ,:i茜5}ftB({|BJ#f+y,{.`׺WwB%rGw άu'UƢҊg,唊KHL=Vﳎ9iW$dp {r9nmy9l)iThs0]01T82X)#$%r Xm(OۙuX.:Adh7M4b sILTZ˃g/'jud%ze-QC~最cI24QV@8(ۦgn1qcJxEaT^pbh PPJhy94wǍ2)׎efEJI9Pǽ:p!}}X=MgYZ**!L'z0E SBexmOh8DsX #䯩hgoeUc| Ble9h/g~p":bƋNВ'Ea?BAڿ 'ZqNywI:gW)tf]Q$9* $ڤ%*-f+A8'U4 _7r3UIQ~ W Elo5UR`z~q~@7~5Wk!ifrJ}8v )i|#MMn.DSCynA|-a>s|n#p4rZy7ܳ4|mH̝# yHˆaˇ1t0cd|Q%`q Ϯډ>jsTFV~Q}PTϧt *a"lb t:__nuK:U'l0_apq O ߿{/ǔw}w[8q30Me$,pcIi'pk迿~ڡk mx_L:.?yɸG2;#b Ώ/x; &\#H pp"dחc,LUVթm+yDaGx1OnT~@x,ub~tMՍ絟FSe `d g"aiBS(Ÿh f=mS}#͛wwښY62Jsܩ9pV`ً̀!֕"P:BK2 8eߛw#otKE+1R\pb5m]/yȘHODGckD7mtgژI>hb{6`!HKp)mH'${ "%eUb4n`gd\g102)i~zw.OOϿ.me)x4Dgɑ"̏k%HC.A\ od+ɞ%۲ݒFlUj6b=H^6(oJCSqrgw;V,ГsO~jmڲئ|T٦se&cX{L!0MTdQ`J5#{5GYZG߼~c!V/V|:N+ͦM*}t P9ip*ϛ!7kPkP4Uu~'>ƫKwe5OIMA2Z[Fd1xǓDG2њΙ8QZI笑NR׬\7"mږDZEyL,~h,MF jWϴ'Ԛ@ߏFw G7jdycS-}e1^QxfԔ/vBP1fzr߷)}M- o9lw;@^Y+N֫lQ ?n|9Vy|V8)mYXTq[@D+q;J{tb]%'ڧ sFH j&~'y49PΘ}RAJ,|j!4)e-w)EB1xDLD4N(jUgG}LC4}ԥa[T(B\ #[AZi!0Yh%G2LT,U] q%9rM+ ;"JxHEIkK~+ j5q/tZ/63i i'cq16Ot1Jj}&'6#hc!AׯNsr;z.׶T|5{'^?{䨼Cw "3 Q@ςY"ntPҳ`a*OD[cU`⭭YrAГ VS9o3h9W3Tmd&wqzfƾX(+cXx#QyUa=' }uyw ?O?7 J)tJHmBX2I+!H&Z#Sim"6`(DRؔĐ4h#t;҅Rf8xG<Ԯ6;Em0`7x.ȑ+ZL /3jD܅W-a Ria0\ U0+m|(Y4dbT) aML,h c*hŬ" T'}e<&vHRˡ>|.=r5Uq$͉`7dF:A7x? ܟNh&+I4-/щI6%j(1эN7ܔhBcf|S-WR^C %%Q%`ĸYb"H:5Q![Bnb"R'̥gڽX,/6ώ.FSp0-]j(gZIs'*~HƀFkduE•%!f Dd&W0@p&q4闯G_rO'2#Jk}zrqP @{Y`f.S&RAjܔD6E&Y69֭rfa]Z]Wq|7Sds"dOϓmH] 棁uK2/7keE5Hs Y,o ]av"޵z*M~z3}G?U83]Lg_$ iϺ۽9 f9٫sso_Ȱ&8P>/Kw[tt,&Y,* cPL7Dyc66@~: (# ©]G["`.˘x@^[2jɒkB$ ^ 0?8q/B~9;^ ۇ~η:|lOClkkm};U<څQxg8 DY/5 h(!䬹B(,Jڪ& q' &eL-)Uvjl_7g"^{v;q1HNѥ+4Ke.o* c)HZxĆ6%9 djT͑خ#6$%JeZGSD9$xY.TԢy ʑ?H}+4Ȣ H!ޟR@  8BFu,1"?3_,%_Zi< Mؕ 'eszWtc'ն1YH-Mt% E#zM'Is<}-#h[b(hNDb>H֣蟥Zu*!+]a){&Q!;/i)l%JVYG- 6\{!m]lu^\x^x2$4rW]tH>lJ7뫧]#?x ge~S^uY;v]/9쭐EOi.tͥ/jK,-9rruC/i8'-޴ll; \%߽M;_󝖛C1D{̻*-e[:^Z"nmLmjt{S *sZBmr}Nz֠b3M]'0]]3⫆6Sɼ#w1 Rog9WUP]3FY Z@UHiad64 3IøQxA&'q*x gU"5~#ŅYmΛrFpa~:gpLi}K.@1_b:e8F2XY~J'!o0"hݔ@j]2UͶonﯺlK= M:%0 \*a d˳VXR.D8"0  "m&& -Hcc1D-Ȋ'C`< 3ԻRjyA9n,zNL>;=-=[bqDŃsZ3OR9A!tX3m‘f0)Cw7j:3-=3 [Co`4 {@K'O 6Zo<2QS'Mq2v8yax&kqQ1U`Ma0#-G5RjJm%ڪ5gS5+A &wښ9npR / xataa3ZxPX.p4H/Iu-uCgR&8%Cp?U{+L p%$w xa,ѿƶ$}^JtPY \O\&YtHbWv'g=csm#_!`M^]ę d'd)˒VG[+Gk̄e)AsL CE$9^؎a_;֖*D}ut+=j7ok׽'lԦCiņͲ>b\8xWU8g%phMnuKiٔ*KW^B=蹌G+OOUKzƷ:r9ޚ޷n~?}'gC=7%$ȑ8ȱ't ;? /!'2\6$7灷١첺ک SMf5{xr#197\UT;} ~'?+<`|< x?]gJݹɇjx?>]Q^LSa-#Κ'8݋523lˍ֨~@/߬]tEZT+5]CQ:1j{]ݞ_,j-e W_bC.)v__潆w/JZKWoY5`.Ǣ4HlxwdD[w^߿iXImZ~oQڏ-ҿ>:o:of/[ݛkgX%`3 N?wWn=Zz@7 efqQr7 Mwz~ʄ]uL7\s7m_ݜnn`t'][\b5銀AU+U]5t5] u '"\h55p3jw4ͦfe`\as)ewz|*sZϗ9fT]۴dizb3 ZUv:c[ycCAsgܜ8cp㜒[t(5Hj'pe5D 0~Q)s8ڊufHNCtUĎU/Nb_դ;!b=OF{QJtЕtos%q ף+]tCQ*6j ,銀AW+{vr7Zm+lO2 tE"`˪:V?luE\O@Ygj(GOCkUkk >t]5nu4-_]!x졁uE|E+0pu5CDk?NC`Փhe0;'"\SMtEv D`uE۳Ҙ=qIS')cޞKDN.A&gΤ :C5^A7 rC9 "1npUSFS7\[ͨ :3t]9J0]^b %݀eߺ ='@v==Fi6DvЕtosc ԣ+{Env"J']PWX"]ppEWH+ ]WD9y']=ePتjtE֢+U\ ]WD)줫 k++m݉V+>t]:cԕJV+UjтRM1J[˴HW̟`icpjj3t]IW#ԕqLT]!}YNT]!moK:QF++m,K ^.g,e4x:^~l֥jֺMM}:U2r7(TԀwȕRgLp -NPn'YU),yGt;Q{r)3:[U>stEx"Jg&]}7NL䝀 WՋ(0rOP"]!Nn=D;Dz6 Еޱ)Li uϻڞRQ-Ƣ;JOڷU+]!.0UV(M\tEk@z+QV(t$ڶg_Wx5BvuEÏĤ {tOtE}oՍ?vELd@EB`-ꉮWV]CQj=j^ UPX8<_ivo,_}_|~}3>\\ͯ+#il%2Κ`gLQ\]\C;JIʪMuв11f_\PviiRķgNի@:k~~ ճޮn_hoOorzo |_s,V2qkQ<9^trm ༒"+5LI <)gm Ɖp%K.(%ḮI"ġ~N6WBb!g_>5 nHS K :d 6x rXfxBVZÄ}LZ[] kI2\`13˒/; 6c8'kCW]bD-TL(ڇ@EeDJT6T3c9&iIl>ҷVx1c5!4֌!fQQ;<72DQJ a!l]{6$Ѣ96fp 儜' Z !qeר U MRIak [^ VᕿK`!fhkJtx0ˬUdX\yX˸3Nc+DqjD9`Oi7"3k\D%%jnP.`+hI6pQg+'D0>kvVNWUP@(s,$zYβ5J8`K6;xmooϵxKR"h,0N%M0!QԂ% K"AJuB9Å`K'N9<9 ә_I.rWL0[>me 8wGgn&iRƻK }ۡa2pew11C+'Rr41a0 X% \ƈg-h^c T02o84Gm#QrRa[U:D̢d,C<1ylxm&a,^MzTPąj $wEc]'$؄qnUolSǻ Z8M<V^x1Hu@KyY=PdpҫhR0uA"\Y LuWAyA3 ?XmAԠBG cA1J#H*HXeV**V:a@) /. D,/xT;7ⶨ t(X:;(*@DXw%,m))& ɘ `| v_nқ̃uP'SaA[wmHuE+pmpe,`;KnCpC*f7]H_GbT@mnk)d2bz:&oCUQs;fL d:՘P%l@} ڜ׃bFP4F,%brFyN PHݔ`.m6Sg$0,QM ;#RF WnP\- 1b:1Ϣb-C DYlls MC:hϢ;K4YF%Gj3KojPRj2+{уEX hߠtf&=G HH/6ۤV t.Y҆u0?XqƜ[tZYx&Mv1M6LbY"MZ`8FL-m~d(h;5,Fw 5$QhS9%  r> ehh g%<%*`:39P.+dshkh+]r).*PP<* GTi'O H5@oتWɰ$>^muErr\ \1DNa};?& UA3mgGql\Ά UۤhK,`[HX߷G>^~u r.kyTR_m?J#F?U}3]A,90h}D;ɤ~WlhuK7Ưq.xv 0 c޵M7-],W3.kh>m+]/lQ[ZNݔz>z?};9Ck݄t90;Y?ZϕF]{.?!pkKր}97e_ ъgo'P$x"o~gVwmks'hz+Ǻɟ ,K`eZJJnues[D}2oh䷒+y /C\ߢd=oNSsx^՟,Fˋ[p>qD~m=c{ovmg;ʖNmup;¨\rۮUmO'E4䮤2Ԣk.tu!ZJ~=4:˂?S C刺Lg僚_N쿯_%FOݯ^͋$_[5i{/0^omB߽\f6,]vw:^ބŶNR( xqvWB_I\isgs4U&:NV_os1)}{uj^/1Loܝ_Kq[:w~:\hq=m {?m W[.!=ImOj{Rۓڞ'=ImOj{Rۓڞ'=ImOj{Rۓڞ'=ImOj{Rۓڞ'=ImOj{Rۓڞ'=ImOj{RۓKU[ԋR\#`r~^/"hzk4ӤXo6W0-U>(V_z9@bJf*D&׳k ,E=:7ծd >Fmt&$C'=ObyN{Mq3`̶A>N^%UϕE*xTIqQ@]Cb= =2y GGv.fTkz>GIke|q2ES~u_w拞|gS2k7>9IWhLJ=yQkE=ZY& 8? ,h~zF0N[:x3jyUN@_87Oq0,xkS;(<]SF"."."."."."."."."."."."."."."."."."."."."."."?m/Ⱦ srjwhy]5JE]_}Pnp eHm­9 yMR4dKB4Gmɡ?yZlͻ${n]OdXWxq& OQmB0NsF!~ g#H2wIxju5߬xx6ϺT;]܄ꚓ,|_s"ԳU5]YbMz(AER.J(\(n<W`_{q֟`ge?o{닶Z#^WNљK0&֐9L9L9L9L9L9L9L9L9L9L9L9L9L9L9L9L9L9L9L9L9L9L9L|Jx5V_^O~@" <ňF3܀ض"2M/U͟$=I#c^ Vm=+[ 0cOޢK8괴\Tq0 ]lPA;t=Ag7e~98ߜ &ȊP9)f@*1JR =]/w=Xs{ONjCNe윱яi:i]}x;TX-(ϳ1N=_/gyz>~Q& ?ũn;__yrPla؀94 E%9yX1^n(G_l?~Ey;ZwJ8`MS\EZ_Z@[R+N˕\"6 lxUU߻0~7a7QtAۺ;hwPn@i#ڎ@6%+Y?*N^E5/nOiw3iWQl E% l'ƀ5a^=>Vhw]8&o'47[W<)cMY'6{tA m1]蚫Pik8,R\ͣ;#n9,ĕ lֳfuo/>ZExB;W#ͻOn9||YƏ4ߚ<ԻǦ7YU5EJNb ~b:G}\;m媔x"\r\~^r\P`+k6L',M ow7y'*$rۨF_'kQT|'?!δU=#V,$.YkX+rd'6sk9)[cŊ\ꪄG)>!tـO :/?'GՆͺݷs\+޷PͪɂXZfwLǮן!Rko,9DC$9DC$9DC$9DC$9DC$9DC$9DC$9DC$9DC$9DC$9DC$9DC$9DC$9DC$9DC$9DC$9DC$9DC$9DC$9DC$9DC$9DC$9DC`=rr=vx@h?n]){]'>VI<ӬR~2t)S~ :f (SOf9 |Y46(Q[|U B I#^=zѻbt=jz:~\ͥ2͝ml(7CXŸ 7rP<=k4wmguy?߭\<%xwm$_!S.ؑ`c[,nH0)L:Wjf8D5-փӚA*kr?GS< 6&9ۇ,-]y694|wPۨ-6anGFj%m_3;nG8l{4c,ϊlg֨uǖZF7̩[ioRIe!Nq\=Ol5v%Ƭv~wi:Ư߃+O9% eΰ GP(E!(zkӘBk=x&U졧zz֐mRJ$eVNRE05.DG i'':n{)(a81OgE=/  z| 8PͰ qY#@)G8A:9XlT8op~u0|*HVÎS3dbV8.c=#^[A38L)!X~W=7 )n0 B2-!,oo, apRÈ&%@CX)͵.d:1C%? kqJtTY4a+Y(_Ƴ)e2nvFyj$T1jf~oJLPFYpn>O7SN3@?W2:kr] vio9^&:ϖ}*5O%|rs´ S:ň=ߚVɃ5momAo}tQCBk_VVצd5빉͘vƷ]r@,lM/t|G+_ YBR{ؚprDs% erʀƑVNȵZ$[eNqUlQ>piԚ:10q{c`{X.Il<4GK(e/&%KFJt9GIMbͲ=Hq%6Mϸ=S޽@4w4qau]& cWXH2D"f SdIo V;g$EU6ցw@RdJ3DPkddR8U Wq唔%3A)LSrAlbb|cRFUR:39b{e9 YO+i'Y8Hֆi'6;T̃L` -ђ¦LB} D$ٛE?UWu+fZ[ fZPR6ӂ.O?u;r|:`Gx̽R`NNףy㡜o~/KY Qs*F4[4nf)%8KL!SZh nƌ%Hb$&N''(sIKp^Utg_W3lQDϥU!`>S^k12pOƁyhW"ZoH[nj#Wz}+ ` +v#`}7H%>1³^Ǖ^e. 8:2'aU"W"[ymd>u2mY\Xoz[z .)5)pZ)CXB8K:3Dh U[k4O[o}⾁&ݍDQ 1\bj4omݶ{ߛzd)FYz-G̏6QS1ފ3Jk^ 58ᓟ S1'  /Hv;M/`J˟e2u\iˤkI^Ԓ)䏬dާ6Yw]4ݻZ>"̈́l_66 mnZc {3iaaDN }] ~n>}_G~!wqŃZ4ܞ ǟ4hSw40bIju~68Nbpy:hz?z =VUorYnTc53̣(𒤤)JX*e jNDӖZ ?Ho#5q8^$" vV. CJg/4tJ.%&fڔ"18-"QIVg ~1u;˝ gS"Ĕȝ}J:H~=>0˔0Jx$T@= Q512rgV:0lt,TD[FTd F@6\ 6᧘ 4$AI_$7FPe&VbWt9ɃNSdIuudD['rOϧ7a$o7o2y{is3ռXx[/(O胗KAFfQ &6% lXN+m4lOpetXZE2&fN@O &cEY8;Jp@42g?2UeZb!/£/hw$4 uf0ͺ'Phy4Ϳq&`åJ t&/-'nfE<*e"6\0 t\d{8C`!)ᕈY_83a* Ea.&~q\s1OCQ[Fmѣvo fVx#T.cFd)4Y.:gl32(qPTo$!eHxс' X %f0 >p>fI:i a1qa/RdOO8eaD="H()$9+4… 'V | `Xa6D)v4*IJS`qGH*)y2!'̈́NS00" uW \-BR*9Ua\T=.zZF"5NEbxULFRKKi!iLepT8ua<ԇ lOæ}cGodWJMj4M7-D"7 UYL%HU92D))= ۠tR廠ɒ,)iN+N$d՞DaL"Û)^v;^iOuHx,t7<,UeeT9 ឫ 4]qjj\zF\8juS|%m7M4b sILTZ˃g_{S#6Y FmKP9'.`4G2J*lzE*HPL (<>!nAp<J #"LNt`o v2)ZdpC3ޅOZݸp14[YbQ %??˿w@/f[ \`9^bup4h88j -3&0t;&0*|8vfa8 솜]j9fڏfs4U{l(UBw K($6C\Rx6 #4˹fs:7zD/~kW| iYx(o͜Y[>q%ӏ+3 aqOA3CkkksoW?f[ŵ}8ΣU3ܮv-82Ol"?[N(B뺒TWr.ú/che˯1'5> ,.>HB8 cenĻpkCĘ"il_)CQHJ69S]UuW޷@w':uSy'ZWT-tqFN?F>7}1AdqL_`׭F8'~6O PO_~>?=G`}aeBT%!eU Խ*\/6GĽl$Xga# E"꓁\ȁL^T.~k6S;*(W&:X ^?'IKId\=*/Q4@XҞr3aSjM+O 'HJbl2 >bԚq8 bac`fW8[pu̫M-J_S |Ks^_q Hŝ^HhLXk v4itY^0.c uc)%F&JIǂq-L=:r; ԛ*-XT'L2Y:G@4 dX 63?SJu0YN;$B`Шw$-!.'hIQ5;+SFqz"=wo Zx;Zw~܁ `|t{赲tEgl .޴\<8rBXU`d"x>H=xB.kOSfE8('Rw2(3_37@Y<ӚyA_sE( 6)Y!@m;LHlP.+hIͅԹw=S"\^wޮ,r-E Aׂ&Gu~`24B/5Gk^n`y-=GY`u8`jei oz#k[>2_9 uFj~(*KkDJI ٨+GNjbսFe~ۯBG]OJٰʒJꡯ*H]eK]eq =u ʛ ZuS TBQW(-*Kt^J H]2`UIJ+nRr٪@i lY\E]eieUR /UoP2p8OK(Y͖\]eXY!_wuB{RhGsP`<hk*!h>hR P:vlfZtvRn~~??OS CM:+o0T@ !W?gˠeqeu9o iQ*nS2M;%Y} Ajު=Mv/SH|ű蝝Xxoˌ=ŔAL2.1PbSJkbՂ6.[-ɇO={`QZ|_ERD_Nb=Ɠ_OYP8) N2sW506P"x0X?pʿs\Se7\1dؒ>b=uʶ\ˢ\'HI.X)$r6^w~[;w}據;_SE/[L\~g弬+T!f:w ZyIX_e_GZyڭR]JS3]ZjU+VoG\q_+٠Vf-3776qsS⦺gHK* ߙ ^ { &()|y +1IـV^8xR\1q@4xN5qXQ%85[JHL\@\1fܘkwG^T3w)Jvŀ9PCq) " ]6dHfQX#U<&bj&M- CFD pRns_օ(e#iC .QW&/t<;q@8I*ypRzSK(T'| *882\#hls,*j^ĵs"^1QX{z\W.ӧ4I#.t_4 Um^nim=ٮ8V/>%wƅḠ"^_`%p4`Z,q$óiV dv>9=$SRh ؍)D"[T#0ROıLc(ցûjEs\8sccITRh jeɔJ'b0T\m܍nj8ƕ]b3s8 2zY\x7ߓ 齟o%Z gOЇK*X3f|F7'6)eH@X0Ad3VHQRfia8gN{NW`5 k^M*&w, iFj[CP@lN{wvG˙]v˶^7B;5[z˩Og&f&8㜵J_tmyฦm OIRfY.Ivj]o=&CErY#TcFxI.R$-U1%NvT8w8`0IUv}hhw5YaSF,dY P`,?I`9G=Jc#*)sDKi,G\])erG4z.18㜤?qa,& ,hWs(Ebh(P )K$M 5I嘬w&ΖB\^ "\H`<!hHh蝀уGqLL ;ݲ ;l`Ys6GB^O%KZp{ Dqv1G~ALYc8oNI?l'6HԅJ"fnB%{#=@$:9'|UΝXߏ*7OٰLZgi^ycV8cs!ch^Xk4PBO.c3^q™LCLꬔ)-DMS9JH2ѝKQ _ȶ ,s)*Wlu-KL(3NC(w14fʙ)+Rd0n:vh;.}@pЭ8mYnc4upEصhp1oַ76W..W9%+U e+rHgm-F\ebhYg˭}[Q.^]epw4f_uGl,b6[ovjyyݍϲhy!rۛE6svj}z <K# 9m UVķN|qyzj;U(}DQd¡}dU&XODp$EZJC"\ 2'nsM%pܳ8[X7dbuQ̏284(Aw)˗ C1 {u$ȕ;"+P%Z3O-Hzdӹ .v i[:;o̳?Hp;0슭T{TkT?qp>L)lrX%)] ZލwXn$S<2ܨ`8T"T3ZBĀe)BV 3j⚆~&٣6.ϓ^x}ࣅ{OUZ]6M[Bf>׭X *ㄲDgQ@c3)VNA0I̅ES%Q3>^m\1%;_9v~mڌNߟ$4:&Ѧ’W24I8AD4:I 4.[(!4nCXTPeNB# :oHArΒQ!:̗|4 Ljww=|Gxqu~cDl inn}k+ਖ(3phպXDB*'xeOvM8k9s@j3FR=UU<(hrZSzZ40j- !ZADxN.uVYGX.jJ|yz#n)=k_8Y u6H"DYc2Ϋg{8W]Fgu7ppFpwK64Îo!Y#QdӢm D4k~e\2h@6i@dNq@qڱ~bfZG !+C e7tDܞBR !k#FUqRZ %4 ɚ΄xG $i.n4V wAΙ w\`crŌBKr4mF'F2&4Fr nnT'y<;3LAŇeҟ!kebՃ.WI5U׿|OEݼ#sFl:@&_4^0AE ŔB=qeX\6QN9M$& rl񿱍n<"_~5XpE1 / 8(:(-5-ƙ Exf:~ۤQ3?'.@9$JzGrf+še_f_)Wtv/^^E5/]:;ɸ ( oӼ{zz6y4{uϭ}q?S9e-a/qskw=vQ=BB.R/}3ך*fLI Viվ}6Lr6x']s(ja $u0fIͽ4[ ɳdv&ZC"96u#Bb: 8dKstAZx-j-S[OԽ읿r F3ʍ~2z仫V|shTymR}Oz'.gSuS ?i|56}9lT1pn6{ޑ!l=4{4۠s\}{ ^n◝RNn#.coUf@h룄]5TBӴZLBq̸,օM RϙI, V\z 1oļeՑڲb^;eLD aþݗx)~>s`X*\SZ>yRh)eu"E(Z(Z@ x>vLYŒ٨9;#yXЖ"c˔uĕG\WO hOẆ[lm\ᒭO87s\?HKmUYQP/XR w dž˛&^vSI;1-+qm$LC2 O&Cp[J%'|?۔K=0dh+n\>Q55e|Sj|>ju 22.;SzXx;wP=3 ుw,D1#w'/{6|T(GbEc%e#-zGtaCD991+\ӦHVEdfLu/PsGJڢ5SBqb"׽MdIlؾhap눹U(u~{r<;FSC?}1Ҳ[^[+ ~LQhiiH.#X%&9aPiFv /f/mҧ`#%KTF y:%ԉG] O;am2жXg-B#]U_q`'؜؜Uo 7{ o۹y1~3erp Źߓ>B؎gYu!3߯/<:[-鍣vܣuömßTGR8QElN07m5ގIvQ@=Eǒ^/Gz{Ie )A‹z>׷7K>HA^s4 l Ș|] .Xt|՝H}V{ w`DjιQؐ쁲)x&ofAf3ǐI6A\jIZ3+aʽ&Q}Sd96|1dzͦKK#egP=PuYDE@^j:i6u}G֔*L `7l=ޛSixmh.v`S=A^|X%%Ih[k'}mNdN^äCvێ6cKx!w|"F1yQ#3ixSeD;M Z(E :gDB0*QYíHcEeȝ"U ve@qfЁJ< 1dOBqϔ"ZnU<ӹIT:b ⌂(8 ɠ4ϒOF1O}Z N}aF-F6E.lct]^KHu+,';X|zu9s.)krcn8}'5 TT*oQ>z LkD Q4sD >MiwR`F8uT"/{,R(V3&E]xV Tnbx,UP& Vn{1'\mHApĐGJ A8VFf8=Bb:pLׂQH'R0!N> BOHN|B4$ ]LdwSQ 3Q)61Ќ/F`B|V_qﭲ3-ס<G2O㉱,@h^8+0x#qDJm94p{xb|cڏWwcFS;LW^d({(?U66;n]І K7VrM:Qbz}pR<1;OQɱ# DΥH 䡀S]1cQ0 9GyQ 2H*vlNTvr,c=P>!Bd UPeA)E HHoO&֮EkY&[y._ YY2*({ ??GWߚ/f rNyfPOk;8obqr ϘȁUC+BJE&aqhvf3Yj4CG(Cg)gv6ol6G,L[@7ZOHt6 7{Ƒ !wՏ.d`/8B?%)C[P8HkZ8=5=_U\ihiE(-`T~6Ck噦[ٴ&'Wޔ._^fklqiowS}~1[nsdiu{ݟƣ2skdvb2ZHBVbHUða4fY +`żϳ.l_GNQ\T̘T_uNZf3W1b~&#уMUH8Bd+d Wjt pƁ胗)%&15 N w*x_ʵ ΁%L3@;eHC%" 9cTmrH+幛ϒqK(s:)72(V'qG E0%0;KLl>[Po GbS$gcN3K_L+Tva1$ΈhIա*zI\ejm]\!F\=Cq gG%`hU&WcWZ} Ҡ(HFI\e r,*S+šLRzJ fd{E(2Jk.J֠)%ev0lKOrūE^"hһg9?Kӄ_wϨzO (VhSAV9{g:or_p;9"I-{yEw(%S_) $5(aB"̤E :YW(Xo3ZEAfda>l0]ėZ}4)lMk_~;Y$vr{u{tz=ËSzsꩢR HN6Q2c|v4n05|QC o'V9=dyǻbSR-^M3/жpwV=VG wI!ȡƦ7ZKܻ>@ v6}[Dm 2mFhr}xq:q~-#P1>wCnY:E㮻C dY˵Y^Z 1AxGU w3W$~߄+9 (W=o_.bgƝPB7[9!ԺekO! ^Zw$R+\']cwݭȈi_h3ۭC_ηEqmۣOg{f-.S3|_KޓV៽f-}75{kk=6+~c5! $9ŴEcP`I?]) ŋଐ+\FⰳQ?2ʈQXE%85[JH.VosrŘUjy.fqCEs}-»j gg(.DSdHfQX#y=&bjg-C?YOq $)*j^uolVk9W#j H I pJ4xT|$՛*H\@8 4d hfPBgRXk8W+}.Ƴqmܽkz|x9)X”}~Z۽Z ^(r Q/gT<JR-4$dٻ. ئ2c^uՠ*eH`JLYZhX!UDN@U"͕\99<(\e4jP:l4WxZ Bi-V_YI(eh_wHX}5%c[`2Ӭ.}(xt1i~O?L M㜵J/ 9Y[8iSY3/:ǭZכvPvh(Xjs5<V[\.!H)ZcKnqQ%Sv.r^kqW(U喚_;uym ݲB( zEY>fIґ`yD1ʙ\-=S>F9Si(薎P Z!l˧"EgQd#*K9"Cs5t2t"bz -U1HL45]9C"e`TBr uXnpR9&kVk#r쨑͕%Dip$)e 9k0pLݲL~6x)󟄼$wM^q&\-Mmtn;,pp:ES4oNY78*>lTh D,8?YFz܁Ht~R!e:wbsT{0~hMeO֐;ôyj8D3dU-t]~P{_] v&3N 1Y$k ^/5ЉwIF{FEIO#Be(q"ClPIqe舚x2h>ك»j9m<ܝ/粀pEn@6envmghdp-:x)+Y¼>Y[v+tzmk"76-}0ν~ԸܺmS60>P`[ϟ[֛tfvg[زrn[=/EK-7C^iDJ؈_B;*) !GdJʊ c.(\ԶphA;$SZ>}}1. *lUIjD~gp ?W*BLyI)jj{blnClb`m9LDh Q;ꇂp}jw&>Zqssl#BѠߩ$JQW$Y(ך!|2 h@בqrKʓ񦘬AMN;sJ& 4dsdI$+R=1S?6,l% tE odX g+HŸNz4j+,+l+4+RJFc68 7SDU$ -h AS1IDwjeBPƹIBݚT&3P9SRΰV9w[nJC2Y*iCVCK8_{"M>O}L8U5`A:!-@'q^ j92zrsIDBqf׶fW,UhQfu "^nͮ&pi{"kKu񼀧nw9ɾr1n֠)QHD&- %qDlJ9hjg>@?(j]U>_i<*3ɵekb?~轍#lCݻ{iU,Q-كT*qK˳>~}ݙ[^A|:%%Z^H2HFkF_vsؖ~0^'Yx HVQ4̌ȇW잇^-G=AIYͮ*XV(nj1  HCs8Tߞ@uztRP:Z$+8afxn  sH$wv"ؼ^ T5%ںv(*=ʕ^]G{~0K}WN-^N{۽hz/˖ s[b⥋#Mn34-hW念9^,m嬇B鬡jc\3_ۯHZ_nڠ\)0;_zQWzsX tڤfW׉n"'GvRyo:#aDE2^csR<0y[eφd4mA`{G+թ͖'3, A7L,< qy,XMxg=2id+|.wu\p"={f37_9hgʯMАlF탋9ei%ypJ%l-UZEQʯ8p oӋ}eqF3igx6vt0L_y")3Б,KD `4aL) p`{F2ܱYdzNL|^ NN:M/ulE#RN56i֯eyzą[p+R9c}ݟ;_H竭`Uʟu=s!kTX\V![S5 2v'7sZMp3SET"&󒕎Eї]W?!ǣ-V1i}Px'&g9@AڛJdJ0Zٹ( SVBoN2d :IS Nx{zJtΞvxÅ8>kepwmK(|oǾ TYp X[6M=R}z|v4NGrRG`k**1_rr< -bR"Zt;C(XO|gBYewN51Te+d:gDYV-wV/1:}%=wI!GWcԾ-Dgvۭ#A=.]%,hD{V({` "mkË%E{%ʛ0ywsɉ4'VmvDkl#ߟ~ Yjh~T*4OpJrPf$ʚ?)FΚu9 Ev/2FtAEt7HWE Afqf-?7*w>E[V#LI|h{q6nrUe{B^Ez:1q5zU:LjqՏ5gj]]ԩcT%dw^5m]ߢx/^y+dLLh,C,iޑ>R1&N:"I^a"&ᘋ{D -"#J!x9< ^Q74z#gŃ&mp%jydN(>_aMIBx2fLZ456 Wud!YЬ y! Ȟ@@v4e7 O" -SD eԥ7r6g=&+%WmmɢEAlIt|~:tY ē8agw&kzj2II7x`256#ԏd̦='Q:d"T" XH%tƸ~9MC1&cR) !+썜acCdbܚ̈}ǺQN#4⭇\av6i9',"ʫ5|f  6'lՒmZI\U"=!&РVK,i Tq{ֈ9R+Dҋy}mu%Eӳ^4;Ӌ?ΐ9nb[': eRMmчqǺzp*쎢=m!f$pR 9Vk?RTޣVYܿg;HD~KP@d 6RA |CעoEKw|jOT1OuîYV#f2+tQF#Fΰ$Y #M@d m-'ą` ,MbpE)5y*Eg bBnQD"g__(RINv6Nرi;ouq9eh$KQr9-:.W15MRY !` 82/P(lYdbIqiVntNgWwe٦ƥ@V˜uªdu\W:yW2D` A@#uN|}<6 E{Clϸ^ *y-yْ(],K'@R Z̢6) ܱ|vPK*Q%6G=3$C@9#IB -3ޒ$9Р._AZ˱H@,pz}J7qrǸ帍1;H% )W?j&DѲ\5"U* Mb~fL4뢡TyfjiJ'Ш'ޗ Btf%}ԗ^/n./&O*l$482){XQJ8LLYQ# >e@tih:#sh(: 3덜ru̫ oiN2𥒥PJbT 'XTZ+ce!9xnYPҸD3 zAIR,4 D&XjJ7Yf:FYFdahV6}vBr,hbH.>(r.pBh^G{H*זm yQ\!Hc]Ԛ0}"[KV!opj.dldl== +9n@[ /Π;܁{7܀)'LBZ%Hup&@XRpRĘ 4i=nhy&n@nBH4CֺaI% ^J Ș@Jw*VJεџkcMw&JL`_fiWc W s\"5+l@; w8 bA&>!HKԔ6!$@EkMpRfN{9_3qC[?ݭPr3cusǔ <;sJK."Cυ\˅s!d}p jN^M )NqwQ3]1?=ȋ2Mǽ0!ѡ -V!.OO`)龪x2_VXt|+Ot2@}.XqA5'.kEV.g"ploTg]ֹn%󓲧+x  DE䧔Wz_VtV4p!@JZQ})wJ,X}{`# E`GZ n(ZvTw vˡ/Փ6zڼz^yԕ}u#RW$}u5XJ *R ө7B4G+90\]i>tuRZ:u眲cRWE`#F]qѨ+UN]Au%⠎H]_4q-?u 8xcHxޠZG/ jK]q<u%~O|uU\oR]>e="ux\E\8cH填"|WoQ]i!GP`c qѼ,rsH);[TWK!ƶO 3|>">#Fz!rҒQq =w>Np<҂)5ZVc_TB ! 񨟟ЮoK!k/sF: D&"PKd&y̳Fn[-5rdWBL "L9#9nd:Mjw96j {\=#Z aN&˄|B$ ,FB [3uL Mq>KH[r(XyYroȒ< fs ˵D[2CnI8M鶼7^ػ\7OãiyY3uˢݞt^:vm?Pͮ$m0%qc>Km s`[FEM1Vz11o$`jBjb$C\?F\??^# R|]cusg *j&wUunZ ~66ÛA5Q<#PG6ښy(Q2'R2=iIWҵU|2A"Zd92p*Vp$z'J :^x!^Cߩ,ᳵɖu)`ER$5cNdpCFF^7c,L0#j(n`Z, g|L. SU̡bq'0@s>Y8[gKܱwYbw ' E?uwzg/_^Ofkʉ "Lh3ΐm!AzͣsЅHդH9:i\Fu!0*j͙pt pPYvhky&ED]\`$r\V ̽uRhJ)c.&Gtt9k4Z,3v}%UWۮ.\6L'>@NM[\42T#J'i @gY\9;l˕C\w=Y;Ҿ媢d5(acY9eRD”v hc\?s˕?#lZG\ͮKvi\;]HbVf+>rf+92g aMki\0Ǹ39nզ,wV9Fu$+3fg &'Z3&CFcɳWpkl^`inJĴvMۖ-l\cWt+YFfYR ~MfقH=JqR -S nJ!Lc:\FPȝ^iNښ8[ۭ'<"\id"AACF?Hޣ(INy< lքq>\ORZ} pF~W nQ7=ý&8,6=H}{7i ^>]L?5e""܄pKV܁h`A/w|I> I:2hխyƢ(pNM&b>:` 5;CtJ| n:);Z9vV[H 3֐dQl9B zȻS!Ayyb>5;LPpBNga=nlSh})),bmYsd;m~h0 {MiƣO z'h8YwoQr<^d.^ٜpqJ6*k˪Wcίm-Z&C\u:d2rŬY_u6hu\,MtRJs0PzDf]!$Xgծ5\5\?+.6o3Q!oVnPqcӽVI*/n!c2/ ~kP~:&8:r.h3r38%\Nbs&.XCP8[NEO+S/c֞A B ;4 \f%gJlvF0&錇> 6IV謏 }F9ᔕxn֚8_!zy=XVPpH/aWߟt0 Իfvxvtxu֬jdl|"Agm|`'N&!2"'WB #: vշ;ѽ}Hcv#p:+c3)ï3,):JJ%uLɣYPiA^Y-=tHX-X{>9\sLuq(ucoPpo.skPp#ȝc^ 1ck5Ҋ>$mGځσm6 o'BC:m#I۽yֱIoGְv XeAޗé76!Dɉ`ap";ݫ@zIkAEO*& HVX3|ޏs||~1ۨ mgʺP,hB&u|/Q^N#E#<Ɲb+($hLF$ ,A1D9Ft`XyZ!(Dd$#?s>{xI2Db Ryj݁%6SC4KTNpl^X$ci[Vލe܄#.(EAC9Q( zk!@NxETO4yі9T3hI \J>q)hvBxJnN '@{@q8U5kJF Mm}h\6{卥WUڪuwaޅ kv[p(`hPOEm[v=Y[BPe =/)?s5n1+cBZlbQDPdy F h7Z;ؠ٠j~|9~8v5cc߿.~p:K,dvm8> pMLvt_?)9uiS-dg:g$2~JY,<+JyS?iN?ܔU%~ꫯ߳j"+<*mS`#x?7W㲠ġ4\+*Ku{Gepg5SN6/!>HUKϓ4-=4qHY P0IFɜyzN=/~_Z˟8.<%`]Xꏳd`7ʁ@x1Ri!dN/)6E2ƻqH Zm 7wz1%3(ڴ312u9Yělbr5bf\0v\w(:#wup,b "JU٦Ũܫ4j)l_{tSm6bYmϪgu%^J,s'MkдH:Pc\bfV?}q勏Cn+"I[w#\ ZJk&ybGuȟb)M#nrp(9Ce+b%<}#=tfj`{٩%wsK`m"мEkm#G_E?mG@p%v.Xl`bK9wba˶lYVAݭf5Ǫb=5WLZEhz]2ɬ1ȏŚx0u\q"jCzϦW<³N~ӭ XV-=C4J > :Ɛ][F~*qԏgTiB3WIap'0.!2ӄ>,":NmRsd{2+QW808MO~+M:ȫƒ _Nt|ڼʌV ;U;蚡TJB' ٶrk{b+ %:Tt)ZQ&Kͅ&ZWBYG!} cl?Q2+ȩn[טYB.mQ>¯];M /_cߥݭn[_&]V>zTL נtzr_AL}Lc#ƃpL)Y{OIC.c,$c89B4`TcUI}@&tLڴ/iŨ/9Wǔ]^|uM}MQFJd2, ]V\V`2l览RF oio?HOӴ.+~^B5 fI]ߢMEz У2WڀԶoH"ȊMs˦m==L{Q49HeXQ9IT7lKxJ 7 {C1vpGڅP4b Bth(P`(l(l'( F@&9Mɉ> ZcvIr`q$_( U=&%a@M)yob&sR4YpV^j}'dwyHB4 sQ0^( !i鵌FXɶ#[8 i:EZ܏d]Ajܱ)jʨ-{5;KI +==W$%L l5ks"݅WP@I0K6x6.xT\L!dEH&&H&ő08,Qs9akc 0vX-UeDT="xg1q.SF9+ Ik5_{$ `oPTfoBK?n6"I' $ ,i! 8e(n9k.6.û%⢮Lq(ȍIfȬAd>,+y$ uŮa5Me<4 .y7p'w"QkяO(]=q.Ӷe˿6D\AƖ~|EX#'Z+)\͐@λ}ۭ ܈񩕨bv3ZQSLeO#΃`I%d5TϪy)I11@PpҟIhz<Z|4hYV#g(v("ә+ CV9? iŦ!.uefkYqN똚: PQCam`H2 8$晊 N84#Ê#sWx4|\-tY漰2DQ4( &ɌV`pnH"#AHT$ *-C~LBO> )eK&̙$B$3JGn-/ROQ{-Mf-KlZI"}HRD!a"UT' mX$E1I*^-%oWH.Ʒ H0}ŝ4q5cG=!K`(NYPsFϘH@q'{h}4Csq:G{ ^:,9I\ƒwP(Aim[c>&m{?,'.8p4CnSԵ ^ѾfkQ`r|PCPZ0^X*rCң_Bpp?M^q F'_'mzkbKbU#)r$nx0b0WsYޓ4:ifbYOFG_9<Nz8ĕ v\5g){HX,[$ l=>8:5Ǎ &f O?;7ޖz <|_߿{E+Οh&)H"!0 yC_GM͇k Pgθ#u9nkbmFpJ|(-&YâpA&1?n-fhufbAU8-P6 Y/>۲=eNhYx;[TYHL9%yV"sK%SA眅6>ADR%e skZECڣmg~g2~C8[Ԫ>livfVn'lnx@㒳&{ 5k$}d+#@jA!ۭ} ϊTZǜM ):xd*bt!XC̙1ȣ7UO AX#X/- Y)!2Mg("EU#gZ3h.:&S |Pt`JCn1yq!1mf$;q7TA$)}]Pt"6&-Z`ud7g^DY:"ʒvݢ^wUb:$A8'GaJt 4df+RtYEZ%Uhߓʍ媫a<}5reVhRI@z31.Xp EQsf`4^:$cgik~O kϺZο[vgP|"w"䲇AQ<OawX@RlP|b+|+y3MŐO.GɛoVYzQkvwTz],(mA-.E4o ΂Lf&/|?yc?{gȍ"  6Ⓛ̋ͽ${ʚIWղ5diumd,>g̺BaUkdHMJRܞE껖^g:PwݛӃhoZl}:3E9 gBjuJf?grui"soHm&

R>_m1n?VWN.\i]vijd**o5V9tZNFދ(k+& Tj]{2uDYv`#݂?fN%8yR k|a][r$Xrf!g=u54*;L-iylQSY:(ՎXUiizy<< Г=aj; \ Wۚ^ -)W$l9bΔ+R+q*puRª7L:Y H- 9v\JWHbAb߬j\K㏮X]%4H \Qc\Z%q*6sճ}  6\\'J-;X]%,XK"}1bvVqE*0WKb\\J5bAV W꓍GsM`U'>OjC"N7<%TSN ϓn/Ӻ?} $׃+%n`jq40 /&nP=MN~ FzarK}Zp5LeQJ pA!hUH+mSe9WE.=;j fK7eꏙ"=\W'׫Q])g{3ʋuNayW]^ξxvΕ,->.x,-l5̾<;?^ y;-/CSuM;m15,1 c??0妕DR4K<YaZXSjO@̄9EC!CgnZ8 Z*[ҔP),9VV~ :SO?v3.|v}Նw/`{WѦ{ԞI4cXs|\}5:䇼!t>2hmP i Sj0h<;H{Z:7vNZ*X~)(kBK.->KmxF;}^i lZ}vQ94ׄK{ D?I8x|R|œ2n/QlG>9-v'($&h1[xi-_!PV/|RL,}HFU>ؓK njo`{5C]V 3d>u!=eBB" I侅dL}|:darJƽd]Yzr}1rZ_)ܪ66x9sDhϺø&K=JjJ9-qpն G0=>,z;v\J'#ĕ4zWHs/\\SLtj;XW|pr,W]ZtcWhvXHpױJ,WV1*qj #x+1HZRpj+V̈́#ĕJ+\\K58v\J;EWLj+kľ \`P Xnq*;s'\xGOŇsr ]Uw~ lzb>Z2f]!5XS%RgC#q{Lv½:IvB=zMAq 1/[Nju*0z1nN:_H Qr,WީD7LؖLWG߉/)no|v۳Ujf#9⭝Wey} .SwtiAS(DJaN3Ӻ{lP 8n"oρe֓VF_?]PBooobniȂ\y>≨H6o9HtA)]8$k0(Nje D +,y-o$%*}}.kO9Kz!Kuvr1BxhL]S[ҍV>h%u`mj AhI_cRmC( !DDHB3al$cRQtalH&mJ)AKE \,6ĔF$r҃O֒X6:Јl s怔!rZYCO63yt|Ή(d$oPMuml-FNyp#ŇK6RhS-b$y1@u Y{*Z=×Lfl4@L!f7iRME7AՈ;Hxģ=ṾZӆ$CRXJ!l*ˋ,Bțd"2HLET$ 8WLRUHMl}cZCi6: 2.7 BA#.x [B",~˫IκAF(M4` ־&/ڸFMV:I<6 ߟ%VA#S!7Vk ΉLUI ڢQމ\R@tV \%xOF d@ decp8UTK嵔TU 4QMj=,^S0u$QbI$ZR|!>bM]&eid)H^$(eP{Q9h/3ЄF#*!("fmkkȋ iXA e^s4I oU6e]R ۞˚F $i4EP'4$)":[w:9i裆b;戃%BٽH9iWM⬤9 EQ b8@MeʛXoK YJ(kJC<1ddQ6NutM&." VUybT)g Ն)5u-x *ChNPBNAa9QT  c$zPO q:PUBր#)CDj&dFt~b V#ZPX'GU}Q5E~5EF|6Δ|d`A"=qK=﹒,k.Ւuy.seg XdUJ+j+1w\+AnTzCk@1e0 Da( VCH(PED&Tm=4FȲE!Eڶʃ Ab΂GQ' sGAC\ɾl:sPI!0ά 8P%@ aSzd&{_ b+joSBFlPHq Ls$eAj g (Ez@9J=W:Դ(Hv]]{jF J2H/\\iLc]hUg !ѿ5y-dJcMȲZh! aUGwU+cu>8"hҰA _[ĥ("ҊYo"棊1"_ 0bqvo4zꮚ!0ރ/oj*+3[u&}6}`-$ >:KAy@xc#Tq:Ъd :CJBUe äc Jrk=/sT([ąsAW) >@h3LbyU,C 2]L1:.zO%$0AG*AP:S#fHA@8z,:T% 9A#+ѡ>AuygEg l;T%gA";> VE?G 6ouo{5Ԙ jm!ns)5ڭIҎa<P9@1^ Bٗjw!=f bھky0t-@G!k? F!tJ3%) P?A.jDL/@yPgUA %CʰZً$P>IYkD/BPˉ6\־KMwBzԞEwviEj@e o[6RS饥GoEt6 i Q%0h; /m:hIY`ȖXJmmkzv+;~ e~LX̻:LSA[.!loNN44z:w#ll! f=; Dn-,F5eEZSR98O)O ]j31i2Wãa#[Uf>Xd5n(!/zHmKmA6G=ʍP[^uR"YLwzʠ- R4*YڂDe{!k=GzT [bc-X@|,D4)K:Sp,8`|F F-Lj RTe i;cQGQ1İY;p1-й'GBF*q EͩN댙 3 EuԬU;)JB UGL:/Q1w@-B5gB'еn=D?TKBT|-ւ L]_$p kJ+ (|<+E8P.(-lZk>W&Eυi M F:lrG򨵧BSp*qKچҖ]\< U׈4JCaކ`Qc6f j{^. b!Cn3c+8{u\Ѡ Tml訉ƧQCoٹfol-r9mhtyIOMnĦ/Wg1oymnq>b>VoJV^bIwZ].oswu?||[뭛냭!m{n59]^W@gRFF\s#Oj}M> Ppu5'Zk'J.}@b> }@b> }@b> }@b> }@b> }@bZ}@/YVΊɿPJ~I|ī> }@b> }@b> }@b> }@b> }@b> }@z> ky׹j|@@}@ұ}@Qjb> }@b> }@b> }@b> }@b> }@b> (!n]h(d Ҷb> }@b> }@b> }@b> }@b> }@b> d|@w7+ױ^s9m5V7w[]W]D&Rj%Bi,ۖNk %a-Ӽ_1iA)n@rڤ5Y_fХܪח7#x"/@]/o_#mK-q<4p͟g˟.>}Wۼpӹ̷ W/_3@woR(n:@65 ejSCVL%Em@ii?g'QֆaXgɛt!A {nг? 5j\ҡ@uFuLƇswJmݑ6t~K7 j 6Zh۳n]7G=jPiT/NWke8Og:^8 #3]}KZKa+D-tEh:]JNդP Q"{kF+BRg_Ci ҕV^ Q]`S"6BWʨgInЕBR:E>zj S+B- ҕTG+kQWO^]ʨNFi++OjJ_ ]Z'NWro=t*hUik.|\]۽Cjh_Kg')ZY]}.iۅ1]..6a |BlkL{?aiӿb{۵nM'Mߘ,b66"{QA ̳AF:0]^A3mx!U9 0R*U]`몡+l-tEh}:]N++RWCWWZоq(C`:A֫XS1S j+:W48 $+B (5]$]Y2vכZF5uJt9UVCWWZ>sPzS+05=$"Z)&LW'HWA^4 kk~yvL-#}4~&ܹƜIV n=kW"25GJ 'wR̋;S ]P ]߯M(+ȡg")(Zq~5K!> 8" 5"Z+PNqg+hnEtEn9y"2] ]C4h+PmZІPtDzX]U"6Le൫S+ 5촬f1u"12] ]w?tЃ5"v tvutT&uEnu61bPJVW'IWɽ0ޥ s~Y˨ƨgi ZJ~9V/3dQUCFv?{ڳ ''kS3΍;7 R;=lѡ/oqC!p_:Q8uEF써fpwm|3*Ǡb[nt:Fj4'G)@77.\\#0V^ (yp]j Z1ww>(xۍwovO1]sզ`WfMOY#Uq4]XBiǒԖ^cQƋN{KBo"%#΄Pƙ%n!-Áb0g]:eq/5HCS"e;^{K[zh^M[Ψph(N1{s-z#{2_nV0?wWoouDV9ܻuhkdPZzgG]G޹UQzֲSKl07 PBM5'tJB 5U֏]G~yKMl%)\OZbmyrXo56t]z"(*bn@FmG޹C"Φ{"pDlSpL9F+1wɈғ)?GKuQ~sK% RRi!+* Քq׆Z8B/9#Aswe-PˌpDv,s5Jr{W<>Y?ħ}'K ƌN|nsKk6q'fo}.ͻƛXs䦷qIRG7$nw5%T)|OPy= {@P)|1u5Ep_m#0u5E(-/\*+BTCWWZ*?u"0] ]i~MEWk+B;}"1] ]YUNh (tjr%#\VJ6" j:He^(̭ܼuM\\78mN㠛CUI5VMƚIe*`sjrgi) sxfjQN+:K+nלVeB! J%o OݎrT* 0Z;}*NƩ{R&+d\=W$U+R+E"Jf\ W'\`j+Rk!u\Ԋu3W (/,}2s\WALWRAqhp+3$FrugWA%ɸ 4z-#\v%qr*u{<"Vsz3TlpEr+R:u\JWĕ:֌pE=*ՂFR+@*Qd\ W1Y. ӝU͘ c, ه_oO.GojB&K>kW$\ZmR4q5@\ikup{ج]ZL~T>D\mQvEr+Rizq5\Y'?ɵWqTZKCĕSFHAYtZ|nВvveP:lյWƷj _ƪ1ΡE4UJbi*r6S@bSU + BaS-9ғq > *R\^ڏ>.?^]??4aYd4#`mЗm^&OgFMpr/,tQbf@q9L;;e!/Ow2_yz}bӀMGU6IxPoMv;ӂ`0e% R!u)w v,a+l_Z+Stŷ+j*x㊿(88+PRWu=x*P+\\ԢKW2`ճJzmQ3 ~ W\WA:H%Jykg *>AlpEjmARLq{%>"Vqu>u\iѾϸjgw"FɵWqE*˸ @`\\d+R1u\3WN9$ΰ nDqE*%d\ W*6_qtVeæߜY2L46M<+!{e(/e>"FM1# L :ph௿^8o9mI!Lje&GjUAܷ3ÞǩVfq3N.(r_TJ0WqXXI*VW$W."֤+R!jˌSvD6" 'o=Hq;/*{'i.RRqE*͸ P*&$J6"qUP "y\JJKmd[WA HJ~2H* qe$WAԎ Hmĩ*TJe\ WNp'W ."N!/WNi^U,\\ 6u\J'2+OL+3YtZ@sIʭ"R+.~RwP:k\ֹK}M?o O@rpZ%Jo'o=]Qo'{9NK\82W:W:걮)F >"^Sl"h3+IW He+V+Ri*q5\)#\^"\\WVqE*ɸ PW$(6"qOPvJk |pEr傫:H%*HO~BP)E^xL[Z0 v|\-TАZOR1LW[t k!C./Ns}rZzq*11\\Ǻ{` Wd"2+I4#\`TlpEr+RkE".jRC0)?'WUyQ7?Eϛ}ɐj9U>y^\sdqv_z'r8kx; ?Nȃ6qxn;Wi(ef,˜gpy>93y'`NPv)Sm_V<哕S vW_־P68zhqFʏ O9xzNmn<ӋzH ?0+jkJ ߼H0ĮWT8>;4a_?^m;yu*.Ezimcj)\JRUzLpՅ) TThD#y"jZLZYBXnEac1P[3,%b#}QWjTŭuʨNϖ=X!9[>&LОKcc9b ^lDE˨CDjT:>ġևN΢0q݅?Cwx.N?ng׿=}1(z?jIOs,?~ W9/[hzONzuOYޅ_+|g(r=!zt'ѺQ,ͯΪohC=A;\om2>7k?X]mc<CUA!7ԎQ oVqFˎB^ϯ!O{s\W#bJPy{vkVOZ{ 3ULeg\wvݥiQ sA]^a w'L|p)$ˉd]7{bW;MvrkZq;STr֯YpaAwSvsV =Vt)@8%4"kvoe;63L#rV3L[ avΪzjؤygyX1K$qIO%cE#m~9f61Ns^f=YSoQ6 &qvToX59f4PKS3ghBj_XMRiASǓq[ycɸMOOۊѫb?/ii= ̛fb ܽ˶1r״w">6j‡|e7 -'Ωfٙ- !}y29{nWne${Fn+tpX.I7cd& Q@E 1Ϝ:’vNu|r֞z]gg_ku7YUG?0vo%E=x Z9j6&a]^Jx6=+H9 ;0,_4g=Wi.i^]eG5NFyADH!71M'V'>YU8m:h9+ _\ iblF#20i|^^;u~'1A8==G=5Άӏa(.a{bh|<y.?wb&,NmLj2S5 -~6+oc1i:v6i̞mb0_}ΚmdQ`~SY#(. fSBY5ʂJRPVn6 ;]Wlu ]S^2kNm|}6wYfP̬.B *U3 cD]3l6+FJjrۄ/z_jE*I?(I\}Dp3/>mc]eŝ\ލwԮce[&cxfFhsOoR?_i6<~}c8WYyr$UG{DEg tzXxuv#B8lYX^}g?]}p$pC[f(]B萗poz\y1#ZT 8kO;KgiHQY@B+F˂鉘.䳂^F!.6}"q ,f'FmwUh&=5L|H\ U&(.i^]K\篵@qLl] QG]<><630ue fˆwO}qJ{yZbJscx$uKk·!TQըk'4JQTa!-7+RZ5f=t-M[- ~9=~5Ř0=\ W |TYL.N[Gn8^{t壅N"ٻGrW} ;|x{1IffmJRRLeڙRFLTȏ M,7iݯvmaj w~LI؈SM8A5_}"7LQwEf뭻?+;Bȕ0ʎtX-{Qϗ'ur^D/ާϻo.2h,ީߎ#uQmϾ8$9{1Cإ"]aZ &.cY7d/O<Ev9]z2FٙijQW#CY޺#xL!M܂~|Iv޺gWohb .@k3lr -rޗ~ r>tm^ V◢ &Y={ŴKY6qȂ3 l^g?Pя-M:\G_2Vq޶L8%ML߳nu d.en痣 AVrzKDcγJv(όx:g1&Y[g'\YI‚V睂Ћ RtV_bt^vso^G,Iѩ̎ijL4v@:\_t}طJ ~WRS aX7K hVZaPZVPZRk 鴄jKkj)ųl2GXcr 37l*tFg'ŋcbFby)V+,uŹ_bCK_J=k=i[Y8szI~KDUo*}Z姞Ş-2z.XR/Q{c icx "QAnFy_6So[2(bO|:gXދY{m^Y묽vT7btVF*0?7)9}2>AeEw&}p;?ym:— ;ʺ_B f*M?y[ߣSehv_zD/̏Fƾq9rp -ߋI&6߸e_DЇM:5)6Yxln>7Y/I_}ߌxtj7 [T}v4E"׎eЂ.__|t:R%RMo>]Lј} )Y+e5ğ^2.Rx`A*`^7$uO8xGjh^` i$ -]#bA)zW6Bѩ( U eΤM>b+sK!ں>y)$yE|%;qzma9R\p%%&3\6)aa*r*ͫ\M6?@BJ!kPZY7ƒ+=e 0ƻc|r)08*t,W$4Vkk+nZf;LgQvEJ&괷m/H017ғpӳo_&A<p0 H{Cv}R+pY`!> kϐ ehK6G%:5rneOt  &MTQPD bH+ӱGe,|Q:a-2\(`w*bɟTDG2fjlN"%z=sǐD/5Ƥ4 Rj)4/RqBhr{C4gwxq2}C'& ɤ*%KwWe%piq6_q5BğC| wogstSBXb8Ek_a4%[%`-u[1@D̹JlH'7tx@ L m&~x4[hMKEg $6QfL-~REhQ!Ӄ}<n8mUX댏vޞ8K1ie=vJ[KfY6~̙ehp7>YGSj^"[6^L4K%Z1麊6A \/jc_x4YG7PZ @!(ʫ(fIgzGyD?\~;n;7.Ő"z]q[m5iiC'D2ݾ}ADΞ5,":(:)W8D{}E)&*_2F-p~OZre>e'^r!e$X ڬU2fW[Nv< 0*w'řі73b%Qe4uzt)ι%5 )2QLsb0IDPO3Ayi7y a 'h@x &u0ݵAE9|}A-TKSˉcC4Q{~,%4#wR,) 66n)Z gAS/˃ߚ'7|BGk֝~נɠU̪y5:Pٓ} jd ^wZ4M6\Ȳje5z3!)Uy&DZ4M( *urjwm8e4JL=A`M396o|v->t!RiKUA+-ɨX(D^GvV*Z7a1}N׳]݆@ &oAISFk= ag\RW1QKD검M7pYw "; 2"Z?&bT0C{ĤBJ5Ja`(h?3̙jʩD=h;,{|iYf\øE{WZ#N1j(6Z>Dq3z~b-a&TxhY>sW%7iSNpʏhto(cPI,AiY!ӗ@сgјFGχ(p2++ U bN#.qעT3݁ٻO_^ Jk3PwDNT-WQ'Mw{ C=kK;nrPM{!;{;>JY")|tvF{SB{XCw3>D u J$ex\ؚ(1/ THʪRuew"yHtSܝ|Xى`7dh*k}KS{Z+T 5V5pċ"Sqw&"JY f`d>,O?`lǩ,K^PV 4C"MVR}ǒ;AU *YPv6]=;8ݝW شp>'4Ё,/^>#zU4xvu5&ɚ;U +wmWh=kVr}0=QDb+tf˷~%TxDt9ѹlLz8qoR{ xߜGחxDN{p^ڳ+(S0Z"n>HIy0W( gL ji5H ~x> ;aM7/cv$;:rDǸ腛 Ǟ5Y9\^b5DՈWupg5: eU%X=v=c%0 RƭfS+ Lے=BhҞ53avBꯗkbw%Ψx=kBIʟըL2 뜓o|:SG#⎃M8 +-< -pw"ʨ[~]_r־ߋAa)U1VżQ7UoP)$4g)X_zu8mJ9x'X p/pŠw@B9ځ,6@1yԜ*AgÒrY!HcPTZq@tF_᥺_㺕ymqG$] U z}q<q#;Te._j*]yub;y%OD=}OުKjeP hxߣʆIXc Jassvd'%nP@ # O):"C aH}Ok_B& Q/yXGnf#Qark!>#S &u:6 Jy^MìN,E=@fl=$+4S!<k C6ټ&zZRܐf[bwtVЇR) BZ2T_K dPXyYeTK$`(yneexDl_~qMJ|>{O᭾/U0Y%K*[ ZNs =ֈ&$_wo>Xc]CْK}jC+ָ (BDWM :r1T @xEcZDg!WX"1Z|ҫ/~**ġHHъ1W13F軈|i  C)/4HfZ@zMy hle cX],`8 zG խh!1Dҧ5>xϋYGJ$% M8^̊Bw]DkY"5DZVGgmo ӯ='x_NU:A`xU/ۇhyCU2#M&rXϵy,Jx&7 nEdaa2֚W9\tyɊ mPl_q6 <뭹&_^_t9?ןLq]oh; w╞e.& Ԛq?p}sO%o gHidXv>X4`{5 $7D 2ѯWgB/#wx,N sbo+_54B0uYZPLp"9K03F-/O{N|fBt^{dmmRFIe?Q-[øПњ=k2?/^ҕne0$@#RO(yVxg̊n2[CTDʾ $ R̃ꬫQw}fE3vRW#+1=1"\(R,m&hf>L '#cO$y0(~w)%,:w ^K5[+\jmxz ̧9rZB;4}R$QKzh>Gk Ξq3J=)]ro6AQ!ga2`/*8vBBdOdm>'JͶEJg@#mz8V [@M\&Z5G_Z%;S88$΀dRtKQȩ,kf㣗VLD.z䴢g}qI僶s'b{tZ_uTъH}zбIQ^?Zxe߿2?Cě%BjXs VB]ޘ5IVC$`fʹRegoTJ8NPkԨ?+B,;2U%@˶c6!s-͟Zk 1Ak։ȕuizo=A'P?EU|Z}bU\Rf+UO{fP\YZ}~0,A=/8z >U jyHeUCtC=K~hqV%?.!guìZP㢔 m5EOFVqSP0j 9GK˚bT&,nt_W'^0cĊ |IZ]M%~EM[9^UBNG<^aVp[쀝= ѓX\oo/e0'-6g,{s..2d c|-p^SXX] a,k ,b,Wa|h 'qVi#ܨvo ȗ+8N/m7P -!w8J _m{ F,ah&S]Hy9V+dX#Z 01#VJ,GJj6%@n--_tN+qm! 2;xkCǼ}ˎ?>AŦt'*O /w/_ ~oDuXyU}p1KY-h$CN+a+q12ڇ ll kX~"؜e+-vQy.]Itۏ=J@ &] m[njp& 4b ľr & `]ps\5ī#鏯վ+uK[g,K.KtmXܡy-fAnCG؋cB@(0; λ,1 wҊgՄrرl#gDt:28mlEz]nBzTkLJSh-!uh*~Ky%ϻT,2[X,ˏ?0CuQӻG=*WU9`ZQ\dBAz:ۤc$QlXA V|\DW't2'y.J'u4BXxĻ]`$HHl`G?r KGUc#ӧw/+ZVWE..8NQ+P0'd}R+ReQD|lb|39L~OӼk,wvfngQddlIcLvP'Z G3IrJv/h93{b-Ϲ*,ы>^ɍ'&If(sYU7<(vK5V_mlE2oE#-g"M9T@dewQ@5Ks5ε+Ԣ.߶a_GbQ M8^̨%\`{Ė:,݌4JFE5X<({A*Їv _ΰ+|x R"0&c#߼вixJ"ϓE,Wt5]N?]- _'ӯ_oe:ǤTZ$Zߪ|eOzlqv]Y.LYKe7y{U?ֶ,ma6O(Pe⿷!o9>9["raL%Ab<Bu&[ ᖷRRi 9h,}/@aHcN\$T f fµw,_/e ??~!n~9G>of%/NiTjx5sdk: >;yEP1 &dO+d':_'yWɱ ZxSm] kfwg8ݓ^%(hHw'ԁ`KpR&؜q.SݵDV(K$۽${$ɟԉ c)doSAܺk9-41>ä%@j@0vPPͷ!e-۞wB6I[`SFUn`+JZJE>9 -ޠ2$MA:Q'k?Q}ᶮ3ИX h"wZ\" Po1GU' - H&@:&َ^G . c I\P&ޭٹ Q ܼ0NlCL!ҋwwŀp -nq\4UtNٍB3D'{x5"<<j5e1?L  G{>XE%WGִИY$ c"nonnn EkN ]k-j+Zs|HVh^Wc^SٻFr$W~] X}ݳtâ#Y>dYT˃ԑSL۲ .r#`D0=.@(W֠ %WωMJ9`Av;qp%6mWoyF*kCC$B CS\Y$Ll ťwDVWS8+> oQ; &1-_ԣ+(/"(bci E$cBy/dDp=JV[z)SZ&6h, H"rS3T %)~P2ͨtֳ߬W*nrpw\gK~*R"euQc;i@BG')duv m6l-=Nq& k52vR ߋڀ5o48Y` i6Qay>Q ϑu $PV:%E+!]+&"1xx S9 E e1qЌd۪ѝ11*O="N4S%3'-h6F=QQ4b"W@{QnpْP`S@q2,R"(j^ JY {vB?$GwsVG_n@Ċ}qk;HpY5GPd,ܬ-Pc܋+0Tۇ&F( $=r^j&d mG:tidˊQrzA9EF8Y"*ԁsݺ>//pCw'-xbۇٯ۩éTJܶw)1.a^lb%"Z Xw *!d3NOZfbeV"1W⍸h qS6fD*8L S!l,0K~9tpߍ<"5*5D`WFPȤbmtkT+q馆K)&E-ZĂ  |fvUC[ͳ c` 1Y9>"`%G7n@|'ڪh"VpE#nG0T"JCcIt1!..J L 9X;\׈$Z.]R rkɕ)%_k$/H߅ `5me_}؟L1{ 6^}v۟SX3'?o#o]SVُQHZsDpډO;cK lD#'xD!::"/4zTysBj r|/ŀ ~ngZ{QDžs `5cti U`쬐>w':UWxGo5mgHyA%*$ֈ+Kӎ m=&xŶ[qaKG }ii)PD)AЖzU}_/Y5GcRH.>衆(_>:"&^:dHH(cٶ`2%1Rtui_n'q37w$Zqjq~;z:ڜ"Mx`pM[V*po\2g 3>YXWQ7+AeFJ] HQ8:ã\*Vtx;z6(6W]kxc޽.n'ӑ^ޖ"#/d TGQ:>WZ=:gZJ5uu5PU'hG+Q r/ ) ޵ pR޷iΫxsѠn"EY3=AR7]xTIՈLV2 ['!D9FhێL^`tpV d \ Uщy^nY !y[ЄChw^ǃ<43[<e cw #gi4*ƺJ>h]4_nbrb# @L30,|&-]O}Q j 2?!@ ͤ=O,MjO%=A%n^f%1-Q^*^!G@b$8c0Tpf`VN? WZ(ywKYclJƺ4Lt2gQC.DiP%H(_s[8J[]6VZ597+#݅JK +{ly(LU,;\Dϐ>*-Ln40]߀ps9=M;"v*ȇTKɀ-%gx:KÊ@xm)1X܌QvA x&HPG5 [X4BFDz(iwM&2K*(_vUE3k*)OO_u9fIkN/o#ϰX"&E[U*EiTH0Rj) Ƃ(q#KwN:i"\ԪӥBV`d[5>{AFtm<[х֛>M-HRCw &du[:$iNchgiQTJct/KsԾԿވ9+>N.>)=AcBMa*0V:LO:y<:4ôHV]_nƢGY̽cuD?-(Lu4jλ(>~Mx0:g*0>}ߚ=vFHCY]nS:džu<}YjFçrPӥO{W8WW횚bp7(Lvha@@ تII+*LaJw}2yx )GIsѥgu{˧1˝͟,P{} 1X> )|kcK͓XuVtiFG!j8UQIeDQ#ϙQZ`SusYJ͹>_'>e/دnFPguG.w~$.oܛ;,zWyBzdciCD$7QWaͼ"ErVXQŅNٽrf뽿OO_>Iz*@D?5mS~WgO_#+0Nt77{H6iuYwQ˜W kw㩗h3 *O.}Js]$i2NsO]\3GOyR $hKS,miyp%D#eeGt$<#h^)yPEeeTԡf8])P5M> 9d/"Γ;w(v@ ʝ< CLҲdp6|I?rYnۇ:V ;aY_Y\q.f L4M¯tv@7um7DK! T,V%7F9YM2NNT'EG;פHa:ILg˷+ ĕA< 2@\m_<$Zeo5lf9tW3Lmim0sto_2٩׷dt{98ZQy)V#oII띏8Wz-MCҽ~B.C ;ʴw99NZִsvlp##9K~%IԴ"erqT:q):-5`nSdΠsF, w_m܇҆Oa')ھ[Trr3HAdSxtQMUgX0a"B!Lk 0͜ +|^hu5mč]ŗv5Ui fjš8Pa#W1\9-LP5v;n&ph.fpx,2X])qg+Ԭ5w9MGlAiWs9b-ƽgNЖԜS`CfIq9Gޕ$RNc6zv7aw`)ɢE,[I(QdfIU2#2#7yH솞{]C-jhd&1^NlQk)v"oRlzȏ]PLLm}GֳWXy`/d L|4̊ &Hཥ]yw)Z < E) hB߾7-SV]I.`;nqA `Y@/zX=j] c)J^Mg"{ M;vIl%)]g밷yB-9'SPn؆5^u354l0˶,XV-qx|ڨ]eQIʃ_I)5Yh 1ec슟'b-Ӯ[`7RJ$Ѱ#oI6s_cB4j?< jK߆5v},ܙ[^+MW/ƩQ`Yuj }\ȆYgw0NqEQdž\ݭP[5qAؤ(a<at+Dn<(᤬@IV{VOb2:m0,~}!{p2@JamX%%{b+k\Ypo Z]¢!տrh#㣫Y\'W}kBhLy 3fÊ[Mps>ųby) @}8mR΁1 33;vg+UK'S4?mq5C /^5θ]_aϬeSi9x/`nS.^ 3ΛA*el v*⹍!.O!LuL^2>4aF=r>" aQQ/]4T>|?'o?ۯ|tƊǐ cB@_Fjm|Ok|\~ogo[{gr.s-~sJd}\Mg|'Ks2(d7hƴ_1׷ÿN6ΥbU }W_+Ot<=7s_+ϾWW1Xr}@ <*RMבi>*ɻ ΃*,~΀c[1__j~j8~Cg-dWkCa4Ǡi R˖֌YZbo Sqo2b]!p-pQDYIdD_2`!uOnL6Σ7< i|D30q+չ&t if IcNzKfofkunH-beT&jy-(uxrha \¾bCE/>Re:,X ?\!B_'_&M"kPs5 ㋳؆RFJ(.'3Rӵ)24tj"֝;5Y$$LN>}ksچy5VbhQaWs.0(ȹ>:j✂FUyCN ܀MݩC_WJP)n&9Kt26 TN0Own0uivuv7?:01cYطIJlDǝt@K󐾬z>0\~KCmMz'؟>nC;[)sEpi5Un&ȈR/~ Ag#G?M)-ǣE8h?"bKzQ2Mr kVYR˯8A3y.0Ae 8%A-t>pBG 1ϕ%-tmvn|qu)MohsE86:Gqho_r ev~4PCHlXL90^1*c: 0xH/)o@gRtX\v3t圓BzЧEQmh=%x2`nU&}k@^k kVJtޘw_R%vu &$,>]1imXlt@CtJiAtl쾻:GgM>ʷ7qH/ cЧ/6\Y:n:AoIQ5^m f~Xb{OvrQD+(!:s\ů &?}p`"lͯȝOgsFq~e82~QЏQ# ǯ'cS L'*_V9#cBc n&: \~ϻσeJ'c/=1 Tv~>"??CX Tk.A^̏z'qe3D%PmMp?:GmG(eΙG[@e8JvH_J/ V(ѹ@(.?mju.MW}Ð.WbEEbIz܇[ t7x6(O[\my-!g@fWנL5^EwSɱQ}%ӿ2!U6G k8T-3ha4fї+ҾCDoB4u00'F3j6qGp }=|(YsuWA&*MOڰ.MRR27|?꾝Ӌ贁Sdz5v|2s|͞*Q6lmmQq>N~j0끟`u?\]Z'/?EFs&1G)@P:nj^N ONet!;;5 ĵ7-f)0#^Q*=1$?Na69qѠyv kCs;5SkIw}mjFئr:\3y ~BGGG+FclPcn!KEBRƬE)#4 9XrDxZP\$idV%mT96ac{$Vz#9* N^R,Utj1` j[ib"w Hw֖fHivGϹRassɩ2 sE)x93m q:QڊmaDJ]́}'2 ݒbK ?yG=iӡ^X&3B`p]\ `*g[% ^:%-B0F$"0NGPIAu fHs"OF7f`x74O(AtL1EBc6 x zVqKMБCKr;dyO{⿇{hf6m~>;-6҃?x} M"> l%w@'BC}OhdwI] RNpmE.JDRR{*.m8X of<`[yT4)na\qM{8y ru6@S4&ٷ<~ֺK5|z=j|> )Qxd80Q5j3랁yuCӘgD8|?lEZ[[3bKy;9{t) R-'H<ܻ՗%G+m=O燤n\Y?RlHM(}"6U0 Z.!5N9=VֽW Cx^1R'~>z}^99;ǥrP *Yg @ԤK! g`Aڨbb>isNSzl(qfVZ(o'Qq+.c:ˤ\\y HwU;)6$D$ $JA'ΓH6$f2E$F'k DIjEE.kwnvZDl4aY<*VTmPnӅr.tzmlՊ^+dđEJD% c  `Džq0sb QBA.) ؼLrlrV>rR9aLHMRLbu ,JC Z6I/>3h5%p"`\ hU.%geDA ov% 4hKSrsđrFhT}.C[pʑj-$^$u@&UEf6645ш Z/.Zk*$|:S-0C4Wͭx<W^I" ih;Q` (jrE-h-qlMBFY v!6HcNNK@,tiUe2.c#k5hK:`) cX2PLQQ:8%XR@_ X mDhmC7]} 8I ϧy'ZIfuá"b8eBlT9pi|־J[0L""t:*rhh.T*! 9 d\KN$my'aaہZV4vh\kYqIh6RO$!@rꂀC-%4`w_R!q> NR` D5'$v@ _e_w#Tf`CPiV UjB`: 4g14SQr8GЛsL]̸䆽 I)=[q7vIJGTGn>vprl2+$wWHJڌL0[A(cMZ4ٱ A-\F˨P ty{n #7/ +dh⽼\qZWE2qm98mgDy yq*hA5y-^CLBN }m0sąDH.J!tfiKHD?{WI/Rd~gb`6 !i54݌Uʒ*K*1¨2`1'ܵ ~oc?hjY(}wwpu7L%䳈\zTݶ-x"W"X>z,iMZؿ?\m+cY!tNT? Oy6wI[B>Jɻ <)9į&et&fV1fe[4Ϟigxl.s93J"3 c{vJԳ0  9[vؗhg'̳3)fhN+v8UjvŜ8#{ӊc fv3oΚ~:8Յ/qT]Fʹ͝~d30Ɓ)kt0u2csưS}91ǶQYӿu<}2ocJ!H3rPiOyfX=V_sŒ XfUH[Q3#X|3֮+t({Tc)@+N{S?) WrU8: R*VDnU:+jxg.oVN*).v ;'.e OqTcte^a>_Uk%ʬr4FUDZ}S&c 9䵘Pلz6H4nZa9F}6_qEx򝣠|tBU4}XYŊE0D@='ǵ;\QBbl$<@RyJ%]TҵJ%]T-V*C=&G`iL%*:P+[d|dcŘOuMAikhNQpjjtRVUVX iyߺZˌ*m9Vpbv]<&f}'$\{g%]2_Kh{U,L{j:m5a馀?ឭwT?I>k3թʼ{~-ob}Y1 5o J8Aݪc 9&T-8AhN>͜%sk9pt[!â_ŐZkpn967J7gk0vlӬ'0h"_m J"gO|Lh']@D%h;Q`БfEDYT7}98FQ|갍KgZֆ9fŬz0A Oo7>r'3}Q(P| 306_L:Akj7Re-Ku 49-mTxڲTj_dym`uJn 7 ZBvʕhYdU),㑫-X+%"$!3g)#B&h@}(ÑEn M9L4 D.vF XGSGcY1zrVvsOtoTQCSQEJdTJIJW hٙMDH:{SM|ںMБRh)sUDt7ˉ$QC. v@>hόrA\XiӜ;B6흪bA,&Ԃ*g'9U:M|ˈѼdMn$?re9;-q'tlZ\uFdejE@bҒTЮ:EZ-$9ϲYdZd7ʴn`'A ڮ3PQ+!6wy.?&`ݩ"_M[NjTvAv}k/:V3h.q~tIaNĞ~.PDG 4InecUm2k=EjI4<|g~FQLa,tX,!ċa{p՗ndzG7m}+NCw\6T e9<ϵGPxV#Yݡ&B]CkY+CoC"a7OYrQ.ORl 2OK+3Aϳcl}8 ^CbEyԳd<*:K!r)b[,ą/Kb[?W9t{P]?s3&t1v 8i*fU3>s61AwCbjk įy;PƯB<jM4j|ۑ@-זU*W檵d|oZTe uzư ^U Vj@ݱQlR '3[$GT&A>WCze[iֵV'!Vg›!}F uW"k|pIn" $>u&< g A-o i(ˑorNY`JyaŮv UtS[X"vn9Ճ~fUXf}w_;] Xq"2MPiRp.5\^vSh-ϰ%O{X[-[Ǫef8ZG k4\ d7e{B<v{ַ^ͽ^uVSZ{a/ۀF@f4֋>R>ż+HDO_'@1sk ro^*,T!(*^NpC ne' )4uO~<9|alb4Yy@zsK©68+5]sfU-'esbkU<{xzۊ˿4/z>{:J8?m)eΞ/_ǃTxȝ͔!FCB;w}O2LuQobW?'cvXs5+cu&/DKp>dp5{RDU.(_Ff!QvgJY8sI>ѩ/Z1zoƝWܨ σʃ+Ƕd˞lקغO˧uZ[:`uuEVWw66H\&!e=KQTzgUPza3RS3-!*m N7۾L9+wY^CE%ZW))s\4VQYZ} >1~طB*y_3zY5) \["Z[(Up coӾUMǵm +PNԢ]jHTO椯%H=ɷfbX-OS.a&K ^9 KG'R Lʱmq:MQI65YժFRAj˾QnQA6Lda$^Bcj<$Wgws6TD9?;T<4 xz&11oJH}VWCF2ѩR?&DzTT>zp0(&0nP}. 9y;fURIf$T` f04)o@wt޽B3}աvq=i׎YlѹX~w_x3݆ahxjk?*1)1ɑzrSPr%7d%nD Ꚉ¢^K D*[F$bsHaԸhX¼y4/RLT31k^_1(T\=L˕׵fRQKmF-ٮÕKqew J_hXC(&q$$9SUEUp6V9ZBg h.b&ZRMcʞ;56Cb~u)G._me 973ZT)QL ?=g^o'z1;FI.EdkThSl5{ƨU(Q?|嬶%$ 'Mݼc|g{noowU~}sn-%%AI1S8!JΐG\͎Zqb X %gq3@pأϣg>ptB]=d.x7 shU /5!'.xdO~nf]eɾteHk'Qa*"v%fE>d6sg**TZ^?L+B(]$D:'M J[R}aKm!jq ` ZoX,b%YZ]JVTˌnKh@.=W*n2hcHmvr2^QMH-ƨ1ތd~I1(^0=΢کGEȅDZGnM.(8Na+|? _"(#;s ՚D(> 'nlWLjX̧wSj2ޙlR|DlrnΕQH3+i]-UJU(CGi9I N" VN!@ʬ+ ԋ ފP|b䣒fj+LWzxqQxзɴv'ȴ:4ΛøWiX'; B7O!ƕ ny:>1D~Wѫ t 㖐IsUt?>u[ 1Lj&݌>6IBF lQbF,eTS6 &dءj,Ն T<SX/N8nǎCH-bM^6͍qT2%orZZ͹Z{L^Bj訆թdާIqEG1Fpq%'d= Vd)D ɬl9K^c͏ tej1S2+W/Gs}=gosB g[a4c>L*NOZŏpSE*Q5lC9`K \IJ \l T\?/t^+o%e[}6s.(]=mJg O^f^g?wyq03cN䤝, ^". y@@E 6:X-!?3ZT]KF{qq_,;,*΁` BLk`o~Tr<0yuݼsѶpKɣ 2vs|.LNmiְwBԼ1O >;1=]B/-"pDN]X=O,=ziK̢bXoJnYL1j1#`8+Q@p܃# buYmmTWsA+CyW4܉=nzuD M~7ƴecQ&™69ڬvK1MV2r`ժsHlV1ᱭFSri ދ.`zwꅒǐ F6 J~ʿ8՟iw>|u ao?Rey:u+4M%ǜI j&4t ֚DŽ9%i8e 1ݹqc ?'wB>v$`>Vw=,8Qx@L v?)qP:Hn1ւߗ^%:H-CK;/JM-Uxy= ov=J]>g>p/\j W]WŗV4- -]j.t- HL{.W#ƃy"%zL2 mByq68 |\\K<'st£orvjO#Zvv S^]1*j9VhB >+?:Ñ EX;!VaFbZ >b;vVݎJsQVhBIi+dv?&cc,2+[aѷg Sk}G9jd?8/0KXݔ|DYNˋ4T{roy29M7vDcsn $.zZC,(`oӋ/.T.?mBKQ&c%[>/*z7ZzU:_5rEIFEcU|K i=fo{!aV@"U Ħ`KI/lT}:{7zӛkԳ!>Jm~{nQ.?bl&p{:?~+w>}6%,>/X85~L+Z1l3</!"x]{D_5=Sn!ǯ^P8>{ꦕz˻i@| ֘QLH;uZ̽6MH-K60f  > 0|RAAm"@+tyH+O1;n;L 6&7w犳ݬLg~qH)Ra=.0wBq2NvnW;ݮv&Iۭ")?bRtw=K/>SIڋs I9*O-^Y4u9N"M?H,F}?=e۳/` sFZ [+&[KM16w|MImF]<}%ܤ~]Eiuwݹw˦?c%siG xxiR O=}P~V{sjPg+W\D=@]o[-yXmJ\5Y~XdLu)aǝ_ |xA8b8q1"UUHu|N; UxvC/([<;),ujI=^UZJyD~")c-qXB-Q+Ҷҥ1> ) vaG"ֺ``f!u$T8+Aa6!S{f5R2M6K9](C8_KE*^ |ݘsC(̊GzrJLm=}:0]~? PMRh* NP&rƔ _ʾTnXR% !I cDŽ z%Y |ꂹ @awgmXWP)p?| &ؔڶ@R?!N݌)B_*qH**9=J=cgKmNIlɄhLCޫb(J=P35rUnn[®Ur*b1|ħԫ*ݠɅ֪BgiJTSjMB]^6UbL)[ %WuhB_v>}jz~Zv?|™m=^\?jEsEfZxo^;^S_c n?A!c2S%{dĹo8#nxbcMooF\X#k*62?WͳdHĈx72+nƱsj2Dkhs N&i`ER^FgF%K6V'JvQ^pӋ8~vJZ{6 R^D2̤M_gFJJ-WE #*H&Q4M:)ރlЙJ;3NG7+za(ܑMJC6($v)[%ىr+P"A =$U!֗JGOMAQfVJjwEᠷ8]7y8$ϕR7pT_U-PVFW.}yK}2}oχa[by$ܘ|qjiU+V 9B?K`Z~zsx[5 !‘^ZlAc~9&`̵\/q"0`/ลH!ZRfHXʹjj M-KAc%?^TٰudiQۙ,Ie.P/ҋW3[H;)Ԧ YDlcH$øj`#@?XsIЖsi5\ͻӏi. 62QAWUP B)wJ ۴#`Tpk x2Ę**1̑![cϳz:%Z]zAU*QfԼE4#VNH+씤(d:7 G[f Qt|+޲MTHT`)K]օ~ƫ@KKE]Ehj90wh{Fm{7v}}籯m&>c$Xy<*E/%]/Ǵ?+b|uIKc \'6emQUDV;3Jg1SQ r͙lkQE&ĘO-(n(A}tF;xQ*eд\ *F-k(o};U_MN:toA,Qek %f6VT\\. \#$vRS׫P=Ůճ99JAUx+8nVHTSN7v#^JMŴQ TvN[Z܊t:6x PM%8tr4W]nNIy缭0K*r!vs|ܬ/GR5(|d:64;EgJ( #vtӻ8[?ꢂ}R I6WLUC#z\&6sZ+^:YYIjJջ_ Fe0~;o _|QUz:2Pj9— >=,Asң-`%и@fr$1.o 薐3at<'!w~Y}e޻=0"zS< E욐"&Y! ?fNPEkQ *\&9xHn9F!yݽJtA sTg,y~C;/y1'r䉖<~#k{?n*[(6vTtAEd q+ XEEf?̑0o<(;(,PTQ>u0 .B^h xoz<|#T9t,a>p.SOxys얘Bx MC?[Ks ;&wChu{#uh-|A/2 .7S8 d0x8W 2ޑ1D>~m'ËΊޫ*Kݬ7x]C" (G:Q'C_b-?uܢJ07K~TE9FAm{oBN\8XDB\qRtހVFOz"bfsqҴ~c~lSUvtvrv9NZP}J{uF'`U=sv %mdǤmBjZ!_Z}uG骢$.餴t_ϫ)__W\jU9믤,_iAwgm9IhAuDMn|(`C{}&#9U*^R'WmG+~dYw"B a2N\RU^^7(ODFw?%Kv8YpJJM4WKgow{/YeS:yNdwF^b+pitr"yD;l3Ճ~ 0=x?_V+Vۋؖ:e:W5.qԎ ss[{g}1Ҝ]~+BTeu7s?kYX<.sɪJj+WE=)3lWWĎh9qy SvD Ȼ JpÃ~at [@ބ~XOzۛHjfhr?ˍD\SGy`l > c:]s6WT~G7>\TƷwU$ fdK#3l~ JhIQe IƑ).>_WȖxhCc#.KeLSXHlB6eQ*/RBLc(0X8UIO.SIXv"qWM&͕ [`n f+bs#.MېV{6)Լ xbdBGv'4K7,E9Oug|isv@A+ !c<"I찕YRUE֙"ҚQ4EFSd|IP(惿uS$HW'`$e'52πE􅺪h";hp[Q>se.rVkuv7͜_bIv۟ъFnie7ץ}zK#$Ox x\{mc\ /Y]ClwUi|-M8aKg?0t;IeҚ#fg;zIiTw#!kvSFr찳7Q}㪨騢r T8娉FMt3U)EMu5Qbɰ" MjĜ; 6N`D&Z#߬R7o>)dhňj,k.3`8JmN  d#&гXP m[,$+R wt;*E5/St w]m6a-i+}eI$;It( B^8$/|/ GukE%/6:;LG%,/?(O{B+8SJY$w.)2/ 0#iد'VVbC椯Dq\wdž%:O#uJs (0!:sԙju淦3QWsAN"vmo9MrRX.$ ;bX!Qdjr˞i2yb Px@ǮO&or{Tލ_Ldu§)Q. Ϙ3[+̝ن9L`4x`hPTdɆ;RPNɀEZ@9 XрMީk懲IHl|WOnA֗C~fƑ-Bd@O&~ytH¾yk0R3##/\j.ӯ.T.b@bI2G4u"h7+0 ۈ[.l0|qͲۻA%8Y9NѼ;!wBJh eN_/WCz_eZvQb/rX`D=nYXWh.CV|fvg|{N{cwA+ if9W6vI!v N"kK؊R^Me~E3 >L$N[_hEB'&W dҰ $=7O%Ẓy嚪};IȃܱN4H{~+SV?>ܵ[5[x"}yr ΨFh%WttCTNwqjU QM7dA1p$[%r99i$2s"G[mTePEj.|}|M_=a+t=bҾFyVh܂ӦTg#rKPšԆ hOۦv ~jmQkݮ_Z\8Mr8!ٵGuKjUK(!6TFUHOw GuDk7SӚVO%s3}gyW􏏶~Fz.@n-&HiSbxR sD݁Z)B5Ⱦ&1 ]lFy`I<"w$-Ik|&aH_):W&$mmNܳe$7@4)>%Q8a,ZBaO/7 4:o%&I"-2E`%'|BJZxi⟻EZw~Y h5ƊgtOQ㵮ʃ;r˫ zS?>t6yq? =9}+#ʄ&Yz xbd:NްT t&z){yÜ\ꗫȷ]|VbBڑ$;PwMxi #ѓQ"W*wM#X!pwR:cbonA>4G΍Xa.<1UѲjE٘-*VY]1 ֎ 1c]aHeo ":)'nN1ꍇ@6V 6&Ӗϗ~¢p>j> 6mi1YQ Cby{ExoBCDlՎ'𹦪џq8;tM^ B|W#ތRkvƖQBu鎪|Mi6f;VhURҦk޺fŒMho;nʷfZeE dbV6s1Z`V:s$ 049ETeɸ3{6Շ1'VI>/?͟kz39 z5W42RP_ Ē 0M>ȚӎI&l·CyFk*"}7^Ys6AB8=iξu?seDj|JPK\d s0P)4Lsf?pN;m[zt5xTVZ(֡p.P[!ztBQs&Kz\. ?)H[)Bz]!F+գm}:Š3DMݮ9ͥT*Dh]|CS:#6:pLX1P+ Gp""6K>Ep,ȎalDv+RU4jbpӿEdW+֑]Y򤐝UW'@ EhQ(T:}R2;>x- \ xL(m}A3K-1͸rִ OReHL|tt "rX2] >Xp{<ݬcȍ i/_\3W#yoA+CkGB$Px`hP̠i US7X0S[8fOi@t`Ff4F:Fp!V4+ 2pzUHIL⶘SvB)ݤiFՖ]ZKYOw[m3s!,c,TIVs/Y4DL|Jq(0H9MSڛ}>=o" X겐]*˰$T {ѝPw EXI%-JC;BT>rkhGla^5L+ԎWͣJv M-ɼ8լ<0_WK+k)ICCɀYz? `*`\WJC.Ɋ%w.)2/|?Y>- c 3ʹt<.I!$p@r YRj~../gP 7u%DߔEJT7'Yr61-,r:OY~s$:<ĶD ӓ ! $^ZnEQY?~[#dbq^O 缔@-4[4Ֆ!)lQ'M<υt \Ss8gt9 w+a{NrDҞ,^+?|@.ՏϒQlhdDM#veit[w¸lƴq `/YLIV8  GY 894{xxۛ-ֿY{gk 4c%nH na~\bo!0-ctRlc*4pJN Re0Ԩ4h#kmXE,2Rwڀ{gFlVd'>[=$Mi &9"-I">Zf­>5:PnO>nun_݁#=tX;HkM}7rXqR-$ǯ<\4>EmȐ>ݐ緧Vz &Jd+)@S%n9{evF6/Xa?v$#P]ʅ/XEqTȴp*{206=@X3l865ՂD@RG /`s|[358me@O1x DAkG8C{.BƧiZK>rst9T@lah8 f։k|AŽh0Cp¦AX G>36Ƨ' > w%|徜}<"HO7)c;/ SN'Q]鸾w{zyOǧݝ=(X2\g?)681=8v8fFs9 X Pb \ @>ƓvdL݄[/#$Fًg{4@˻} ׂs`ݾ"A(4+m!C$j_!ZRK]R+->5Z ͣQ{<EǶŐֶZh; bPjں`u%dE#X1c,(=h6ZI9t7|O 5-k5Ӿ bI@IWN 艾9M%ZgKBH_@]:* [R h̆^d7g߲h}4 U#=fQ.oT W1D%! O YxJg99&!=F %Osˁ9Vڤ |"Sr̂lFeVF3me`jkf pN82$BL+ ]*͋\6 .VHEOΤ@/En>t0:2Ǒ{Ÿ#MQFO2H!ビG\k XnK^ObC4-a4YTE9"!FU*E&Dj&) w6;N&!>ljĬP%yp˒HҐ(.I``ɮ׸v31_b(rV0Ir+c>H&a+PlMi įE e> gJUbQ)KCm`iTPP OG V In/I<^IX_P l/D) G)#󽘃@?F//k.!/[J ti;EA0):Ļ.dap>sѵv6ߺL#{igtgpUzv[Jȍlk>BtOm0Ƀ^sQ17+FU(ϓṣ%}~{w?6}M1ʫc?{.mӨ]ƺ/Lk#O|@qʀR5Vm9'o0:PDUz(|ŒІOд\P,4<X3`:9Оlt}I5$`Hם@pFJj8a  rl&m8ir;96HȆ)4lԖoo$8pȨv&6A33=ѐY=I nQ4cMSzqTJf%KFO&lեdIr}_ӣCaG%CzXo?]+e$WatJ#9@m{|f>F3aלC31dL -g"IpE #Y2HbS! sTA hmԎ>Kl9{??)UآzԅY?]}?ǵA_ ;W-ùIx?O+%?'|yqWt\}[Z\KV5>HȿoZX *id Z]ղ̣me|04R>Ԥ>*h&' \2X„뙏.4kwLTb%m\ʹYϕ*4ؙPR_2L+f0m`agrYvV*5+5 '@'r^gfbXrVdP4=YgU21 nmCK[9ҜjiލZB>m> .69?GZ7[kMa'''&apUl*s9^v@$BFqGumOyhiwM2؜d T:rIrTob2Zȁ[.ir\s#p2_SovUa *o^kߗJ2響@(-'iAKju.!sr\ǁuC5~LӺ"ŏ4RrHV:(T1;ҝTl-"xFiA}'K|>ҷZ F1 pUA'13J,$@8T@U\1#8KQ mBq8g/FЊDKqZ%9qx B*}YB|c;/&FqZI;rV⿜j/?9E?`LWEiOo}n랅v :ynGL޳59m%dzR(Qm= J m7ӚUϗiDNX/gZHFl!JJ!+'T[.lNDpʲ$Z!L%URTIYT&16QҙtL.k9KaClHdni^ Nُ "݆DҼF-A/eBt@1vi6j*S7xH?<!w@&zT:X!IaMoh 4ؐc YMFH"jS!a 2A&(lݧm5)JR3@V3A:uFayW!gK='j(G<#ȼYO^`̤@`H1,gDf2ԩkh](tqZAJzA![{hCLx^$-&tgI8iͳ9EK *oSحc0~e0wPz0s]eW\^tGi˯!?3.|{wujyԣ<ƷQh.nA<=Xr>- jSvSeQӯ2a2('t|۔[[,%U;'" L5;]{m;zd\Yr/1$Gr*'dG(ipGٻq%WT~ݚcʸ_Tl9U>T٩I*E[҈L*}$%Ql[8H@׍Fwѝ҆Afi='d G_xd}M%, ,Kz.ޞ<4)#xℜ6\3YKA;P~:4pt~|wq!4T߹٫R}g\BpBP YN< ˁ/~O,mJX9Q,XAvM vlװv5vM#>+^v+h+7WW.ӣŧH 'e f|fZ]if>>4bAm#x;)[IF|+;);MEEv%Qv-ptN.{,{,`ŻY6^լo}&I_fGr'Ty}=Y뫬-E1f| )L[z1J{C=I_.hl?4#;Y]彚yRN:*/nZ\y'W5wC8_Ɖ+nQ 7gf:kDmz6߹ pܔ({ߞqLhD[qy0,%b1RǵFU0azC=w$խͽf`H=dDfS! w; 啂!KXb1@h6rdVq;m|j<`;^[-OAxG/(Qy073y z%*>UWVE5J_gЪFa*F)"g#&!$ 0ҡ5tcw|HiR=iXk4}`XvH**v۰ƻ-ye+v#>֍ 0q|-.@ 6.X Q"1-#Ŏ(,L݌HBT 8Њ[Da$R$j%ٶ.J$C(q\Z\D0R8v%H[*-CL,R/m!1I>_pJE#^C]s1f 꾤$YZ@"Jԁ̣kFp0YteLZ]VyiNZ/qfԜ9$Q$3(*yT͚xPDR:Zuo5SVr< x #w+d儦&I"ya#f9tyck-&"m5*5s mWWBe! 0o9&c9zyfYRg$wmFq!G`εݭ ݚp3>ӄNaXnN)&4֎)]c <˥+]T t6bz[39! U^8N)A&4bđJcY%4g7ځW8\Ӿam#q;[ F|й@V ̂}%i7°f[Ujw(8K;8|,0\l4NSB#,4Q*d`jZy2|D2Jz@趟iC q,iGP}3l>DSr[nSv)y@v簫`oJ2 n9.ohe0`2tyBl1%5k-C^,&rc)vJ0^./eC@ww-Rs_>ymhT>6>3l =Źa =l =יG seQS9 gu96u9B.oƱ]8̟Iم{0MJ,4Bi=by&v`'ee.$~,:h+ZSUձGjpZ?Z]@$rׅr42ww<ĉ[[uJֱ ٟ 9PņDY T%X9cw0%it$b 4\c@$)GK0c.bm)O,!DZ5N6 Dal:8Љx\$6[LR^^NpSf(0IMA!Yn_> Mhys&R&,J܈I,;.MӀȆ SpfOIocp?קjv1>wiۻD'E IG2?RnƓoy&oʛ).Y>t:`AzK_zRGN (K8jUwER9^j.`?L p4/{@3w9mGztR#qgCy|8|1=FؓS丬ؙg㣲scգso.'Gކ{Ep:8|=Π? *Wl0u5Z5:~:ΖN%Я//qXyH{sN‡=̟`I÷WI~;O|i'_'kHE=(q]0X^S%_MV:uv5*czU.v).pM~uqloytx.6^\\,~$3Ews:~&'MJ/YO0h'dOL&/iڏ÷3H3X_ D+>Sfx_\wyv6tŅy7gc&+zX|d'(7uE_.'Qixd#@}ױH9؀Z m޽0`a<^rbf+Y)Go6i/`vFΔJa/n{#TNK6*PW}"$GdFKJh FYzyngT`('CBj^{gȨ3<`}ޤsSֆV:V\D)^W`DiA6QnigtQgugt&JghlTe,Dž&Q+1%P 4:%`4Y)ᲳP67R ai # A`HVZ?4)x~wh wR)[(_@tJ}wJ}}-7hm) :bQLc(JBVhwZ}8gʉ~&}̵bɣI-$QD d,֘jgULӐD-'ih KKANž/)QJ5AD;m4DhaQ+ &9G;3Zm]UnE + A}%G .BsJI,B(Rkօ"9,Ce{ W<=l)n QYjXW /_^ϛ#W|g k=G~яUhǑIDL`Gs,1M(:ƹTHbt\[5ΨBk7kCyaUc ОGcfoGsA0L[x $zCy Er.Q0Om f-S"9B4e.̰ԬP= rb+5ٕ v%:o?5@)\=]kCAUj av)F;w@c~NG\ އ%ЀJ&$OpGC92>z>Ӟ ~fӽ1rr==t}rA"˩o^z rr|@"QUi~8d8f DBO8*EY;7EM;M}A\ ?#J)aZ88 osz jnBځPЎFy֘CW e#f#ϻ.7\b"@Ыl@rC HL1lD$1!(L(0*1Q"7^"V'˵U DW6YoVΚ@@bOR\vnl7H ŽLy"caS<B;gN2 X"Sh ' .p,Ll &Jg1ȥXjAS8]Tk&N0JM+)&I( amnRXB I1q"Qe& ,6`%/#8({U҂MZ+&FXYUc+XXFTߩp R]ù=ь( :hjM})Aw`SW k>pO4MmJAE-;97E>nf@Wy";"pڂPtDuþe`y/0.3MXO0Rz>6;Ƭ7vC Q=Qf4gѷl; ]a06]yFC[{ak[֘؄=ȶ@w`LG FTBk )4L@vS &t c)Td6L"]o9W}^|_C] n<&;ő=`[nIV?Xbai*&> jD?'whMJl"zJ{@"NpEGB;mNKdCK=k>lEԔ<1Ck*:{#n(ԶQ J)K"$ UZtԜ1T;=gDa*>VJܟS͆J(:6k犃6b6vYOEڬ'L7.#oڈ|dEeP->CaC(p&F^>6U9ϙINB$'\:n[aT鈶ΘuNRq񖌭ދ/ aTerz8A0P32\QSB [FzxS^"1V[+X*65D) Ӂ'dF Lwm@(ժڡ*%5:eJ%. ڧgMfRPT\3H[hW,-SPD['¡TbGy߾=(XTE6Θ/hjyqK¤ZԛGRToB4.MJ,;yD'ytć**lKn$ۃJrܼ蜜b59єAR;r/֟tI@V]E4x`hٕ{|tFf`4nظ;&#:$0i{D1=s= @,RͰ@9*dBm'Bs&ԢLaJX+BwlɁwܻu(IhrJ1}16(InҸcf-*{f#f36xP[k.D*:TF}\Txݺ#H-1ZC5? g}KDu4Q‘aذvwL Ȏ_\GOzUȅ4iXqi2>9/VW̔H)sܢA6eLs!>oX*9g6m2{gZ4{z3ōT3m%HAzzIֻ,Se]JD)B>dsdD Rܼ5I%qi)a p`qS ,xJ#R&3MW3Cg @ ƔQÇj:;&FEf/CM4=s+B=].3&TR"I\.5q;tb'OZou(C XdQy;Lơ,71%H( m0WoXܑph觀M$o'OͨBB+7YoHtjZ5TMS!I6ޭpγB&.n㯅.J|NiN4@5d0~i[@/AmhKP7Wdr{_OTey7$^k4ݗǭJE@Iц14 [YZ0\j .4%%^dVI@ifPnF!u?LC _.MmF,(r,IMW:Lv.Rk떠ĝ.$t{r jNNOa$%TTKt) s ZZ\mRgٜM \H~6Z1!+} 4T>>cr&&m8::\*>\ԑ;*%\)㻕MTp=$swġ[35f8U16N*ɎBG{zM5jVghq 2;FrQܫiBฃg;jt8:7پjኟ;7_'7FمAu13|OЋx8}ը1`&Z o5![^$EWoy*ڨFB'BFZţYI{\Ք_5f %11Xބ2o_VcUƀ˘dh K3x}eΥ%piRz%!s;h~pl[tN;k@xq!f 6s)ۮK`,GmDC?$~qPs8tn>s<\\$ nۈ6J`DD\z{?hr]xizZoyM qy͋CO@_Rn= t ^hNg%Fj͢:ZnzaRʆT *9q_ÇKU\{h/ <#GsA(Rt`$0SAmʑ8 ~#lr.U'YTpw?41^"'>ҔhW}fnBzO ^{ó~fhè~> rz!x /]L D5V'(:Ҕ"o:Gv[сOF_m!JFv7YjOYe ؗ'-6Foo4́e\woxSPN#WT&Ơ]FQF1{#F# wcuf>+p6ĆVks,Qv腙 4xcBzCOx3x"wN=J@34뜎f`"*o--r0UFs  惟%KDɑB'M<Ι=ǏfUlm ԩFYL7.ƗZ2ѯn̡3sU'{Pn+܉[>(ȕe5We"Y -8C;$ ja02=dL Ӊs~}%]rg\ bS%N}vQIgbݕ]'ٿY,^]-.9U=Ϸ~}7Pl&|՝@NK8wuo&Ko }{M2[{\&Hxb' fᛮ !KJ!I.˂DC+Q18ڐ?9ag,713B5sE ̰27!Ja '^A?|TW4,ci\`9IQVR…*s\ڥF"at{փyΥ"TIi~?FD=FD =# Homw}}m_A4DOOnzv n<_kR>) ӷ?u)fFZ߲o?)@Ì2bר/H)#Kf/?\}/ЙBD=~{T0 AkZmkuYf pH4`b:2ǻ 7\@RjyBT34Dv]` oRH{-B*@{}m iجQrx}Kt]1-^_}=!*B*A\~O]P,]ׂm]?305g`CD3wj=XmрdЂNYg)Qmޔ᯼ǿ=0Jiʬn_bǀnPKc>UՊ4'i3ʕCCi3ΡvY`Q :M;M3!Q^\Jə%UJ/cJП̄EYN" Rfzն`gpj~D]}`ҤBȫwZGSFtaKKD3۔gLnQ;Iu``*'B|R^ [PC[gR *.$SH%]`It@]Ju(,4ޕc=rw1h/w@Jɻmiőml咛s" mϓί 6Ofo_Ia.3$#$L,>a:$)J)0|^~.bXeA׏~uwbn}{FqoN8)zV T2!rz./Vݕ7y 3Bv*}KZw11iWP 5'@a#U$F juI"YM]o\9rW lK< H6` ֒%HmxS[KM>˖@e7_:X{{d3Q~X/[O%2XhuX[A>'Op޹5t6(<_ſ) #[6]ٙR {A5Eww'摰&9Dj+$KbW|:jj>qCTKfSh>9͖Y26{ؼzR-%T@_\kb(7zxg sK v<{;=?; sroU$l8⮗_w/k^~S?պaǻQK1}ףGe绀)z< w1r{T^'Rh& .5PVЫ />A/2AS Sj7-8뜿^ `/cj!A=S\:n*ZSoPi#! DɧoehD9!Iz&S8΍>$a$ؐ"M<SltizN@r>$cj'-q7orsJn,uf3jN,8&i ~},.}YsPQxB;}=://ME9@ff}3r2(W6("AP6ȳ;@^ U*C(]S*8V1O:.$9E,Xt-$խG5kV-T.M')xc zw]z{z'ɒjJ(J8 L7˽NAH s4VX, +=C̉3&@&Գ\$RRƴdH*(FL2k 7' 2R^^2~ev ϫ\tsuyrk77^\U-Gi ^ էi&l7O?.M]112EqX\ujPW1>Mu6}ldggS/2l]ܡ9^,戗i*obt]\ _^UIiCv>={)BiYS)CSr @3*Jrr";{HV9C T%aoaw{J/ T͓b# ו{jhfIf"%6W3F`$ۑ* Ǚz N)Dme&Up08z-Ё6-<hQ/*ݿDw£e*XRƒ7 z.8u.UjfuЧxfZ 0lҲv 2 zDSwwOW0+XF,?`D^&^3L+'j \l5S]4* )L)w"F&c(g |ݍ2kNZK̚r?gY3srsJʂyO Ba$2]bHђʨZpPr/ n-&rL"5;Sm=ztV޿2Dfq5ރGU^`XyH4X|4Q&ۑU''t5chZGn^<:.0cﻣY!|"hb:61L7i~U(v:+Ts^jN`JHA/Es <#1CKqPHO<~b QK@A[ GD6x#H(NK? UB ́%OF09{fgզUpZ rb'k( N-44m: C9v)4~j&[-9S;GtBVTҭzҭy,ZS-\-]n4t uj(n"B&ݪ7+g΢[<$$wk׶:5?8f'KkwC5(b>mAڤy,&@1h΂I ʣ&Dfr[%U7A(f'a?姏/-KQ&#=6B V H&-k ʔ Bx ։J*+7aTFPIQDžI^QƼ,7;Vn> <ܦ^zgѮo>`kZ}ZL|{/Wpo>U D[CVHҩSM}GcxTE*I2%S{ IaOnv#[%bS+INI$iy7%} .ݮ;.Sp~OISԃV&ꈢf{{" S*Ӟ! )J/L蔞NSj^6H.K2\<f^n6%=f =F pߛ`ݜûUlrXDC 'RZhΥ{O$5k]pMIZX)M: H "td*!.Z{/\t7G"^j ?6؛*w=}i n.+?,׉XO˸9Gko37`Q [raEM>iOs^ϭWӝ!'maJͧ[%ׅ.u91ۓ|d|<2n.mx Es|DRC4mKz} .~^R’ǟs7j?Pa˓c%6U_94+ *b o]~{ oʼn46Mo1>Yٕ|z6W2#Q3 5&/+VZ,A Bwvp)F w ?1*ѪCrSiz VqJw8рʖѾٔDjݴۡ\XEEL5u*SU3fh`?S2vY`c֐3 Y0;?hjDfw.}Wk0/`reRNheG!x빦o8ANOUl7F8UL^uTWR0STn.WTÁL0UOH Rʶ~9~é;1L I#t:o8Fv˔/s7s~e. 1!t4ȯ?+! ZhA7r 3tkd$>v"5f1;v7cϒ&4n,0sf=nXY<}Ph̑e{yVbN9 =yhCw}`ܦk}i0pAFqRg6$ͽ#g=o=: N2O:ewĬ`(LH@̨GҵLFjY+( -K* Z#‹2I *A*WĴbjTNa7Ug.5}EB^Wf{@9 Ǵ@R2]i}!ϜExYM:N(j1a9Jt3TMUo4W!ϜExJ[;ME j1a9J۔T0thtC9Vz3 jt9]yc6~ӎ@ϬZ't1{B~z/(M+xFQUB՞FŒOT=:S![4#-{)wVRBt#!(, %xqZT2%Yb_jl4'>hO ),U,?0ǜ!otB.}:޹ eC{aʁwK^6ٝ5r`O-jh4u RYtҤYS4^7%쮀 QFNНkܮl٨^D8>]\\_@v75_ȫgŹ; *v7wmG_A%мUd9J8X` dYX],X|ؙ_OOwOwOJUy'5Z#>zc֪3"έ3hZ[P2a*wߦnjE E4G4B[nBV=cJn{ &LK2Z y"%S|}SIv%?Z/A[5P7}ӭk3(.ovGb/8950uN3~ K&we끿\Vnxr9SNׯ&z@K &yv|" oBo_:/Vz)/U<,>{N|궧9_>Mjn WH޼{pOb\HJ3 ϼa39Rf/b[1FD_o빺˓33506A&NӧE'&ƞhxo\:~Y1奰IY8[^y7?wӳCtT#x4mI,?oiXL\we$T@Rs$^> iqݕ# r!/ h}0$vl5:t.Dl}kzJ,wS3 pۄ}'tn완 HJ4ZD>TWh*iw?NOKOPO'BBdIPֻ~"J=Z]m'U~Y뤥]c8˸o P!B"`Oed(gS}@Kr5[Ԧ.K"@Xeʧg_R UjEԐBN`4x*U Uä..Q?ۢt ynSzG/*q[ $ra*f曯k|6]S {/0O)Ycv~ ozlS?%`XI%} ΃I;n㪽6.Zh`q ckT01$[C@H V|_g!dvEyGCAgdh$"}}?]T7G~Pu9Tۜ((n!< )'kKٛC^;*:udNQ^QJA`EєXCCt^˕GͭpA!YR $D˽^ z7!^6"w_ mfd&w_ ۈg5g#˪2-im Mdb-ۉȠFuj7k6}wABjMPeRg)zА9,FK;TsVV֕:#k4v*"AAkS>tXJ~#]jAe=]p K(g}]{^?uPOxJaqY/5„K!9q!744q4e(yM3Q̸Q]Y'%')X1D*CׄSG(&!5丏(u>[:%a >g\,#D[s}ZRM^.B<[!hmMV~$0hP("$O'0r ܞ/ +wIRp'In69rBfK!-g{!I82kab?5o x:5M,Z"ު4ڋ¦.7L6aug4%]l1hKﱉ(oLVbDdа&]1`7 $u;+G>`V[ h熆h?T iy0d^~U)L:] *òTuwK}'D[x pqև\}K&a ȡ8<ݝ%~W `Sa=~cqY/5DѲ BTCru$ւ݊IF#Y(k`=+RI 2<wGjoGkhw67d*!p#>7ȚF({>zB P!C%@ !EQ(B¢$Do13,P=e3sw)҂r#ao?~fT1"_NT򳧞/Ug nf3;ܶ Z ۈlE1ׯ`"@Mg\`auXa 1() 1ʲ;y3tflUEy-ݧγ%=LnVv8vRj l.bx5 ćKMjw6J}y:`c,œaM5'>3!ve'_TbXLs +$l!eQ[Xkfޖ¬u5!$Ԃ>2YL1,+=0@4Yg;*w{STX Jӽܧ-qȸ 'u8ղ܍ݸ.O G$@H^@dͲ2}.+ȓ$'I"OD%\V&Cds.ݎ"x썍$*M$rWeee(KL@\$8QFn@Dʔ@+ kzqBI"JXq Y3(NȲewb~G.!Qh9;q" ˑY" ƒ6*,6rE>F DQj_e 8 ("g2)g+FUNqӝb}K1Y2Z]2pc9<䘖t|8v:##ahY},FYc1ñXJ;Z򛯎`>##'M}*0γ=kS#.0HfbUdQgj !͵h+Uk/n8\d|69nMf[3?u'~p7fm, +à zP u0Iyc:=`J@]F%i : Sq>.Y i$igj*쐀L]g<ݩ^{nM']RV#gŕ\h5R4(o*1 דR L.:*[$w(q0Dh3H}~>5ZAt19 '3H)N2)ByǙfnf\0xђvo6_ATS"Z~3_YFf5M2ŕ {9\^+[\7EJ%jsnb%%rma>'Úh*DbM$$ G)%0qbq1N9lF0";ǂ1^xR__l^:x'tpDaY<86r|GbxEj@)lsU}_{@knL}UbGĦt,7G_eR}Tf1FŶt[}lL"yǺJx)=s>"D7|=ۮs.>{k}{+EC_)4l336[ЪgȰx4S%dV)<3RE !$/rr1lϸL?E_Œ?W'L!uA!|;.>Jy3ϫW_~~9ˏ!Y/9Ըr2jə?w|J]h!p /M cS⪬xûSZv߇ܼ 㷗Ei~\ &gJWM c%=i].[/`M؟?.L(L}5|W1K1)~J xUWϯ.ΧxI9?1H4mbNzose4o49#P1}Wc^1>.)v>;3JxNQ=~ j5 _+h A.g7!ׯ}7@l}{H٪jblBe"PWDžYMqARjv 6tꮥdqfvܸ:!ԍ \ҥ`G,)%,[~7!WIeU+&N)BS7ILʅ28DxaG.us)LwnՀ%ylAއr=y@LGd`&(,D*iB(0hr 6OʳMJ;xᔾ$q2 VC'qpiT8n K0] YY>}xN[jJƌ*M#ԂpHB`T8MF`-zW}lbf(%U4{cx \<m޺y= cv\H]_dkH,ѓЀ 7~ѿTJW aE̯WO1Z (@U l P@I\rFOP^eP^` ťAܔ04}M<1U(CA I.] j9H>Ld)2@T0P].KMKт_8&[#LJ^|.:kUzejAK1tY(8nuqi g0rf*Qh,D0JS3A{oCibmh-! qХq=db\T%} B|vmDQ&:ji|qyՃpBaut Bl/&ۋbUWMZ!Q.eʨ=2Uw1!Jd;*N!慫9ml6!Aaz:vkłf U\ͪP^[/DF}oNW{ QTv7AŇr60\&{we s( _;~f|a?GkC,͢װ+TVU,(c>3HIS *gL1a 2FeaVcpB0RI|t8Cb} H\@t 5>Qet:";ZQ5{;K<_lO[ /d*ti7;T7ջ'&Pn aUqj ⑁g:PmpC"2Py%Z;i\*:[_Tg\ro=VL*ЗvNU-*$-eЊjj !jjL7ak)U;qqEp#}YtH.P2Tܾ2S.kx݀ZpR:B_Wjtxou{r|7M(^Cu'dn1>;ǪG^VJN&wZ= ,zĪME-=*gƼwnԌq!'2;:Aa LV w=\h&m82덓:tޔ ynF01E匼Kӧ4v1Ο  |zڰr=#!_6)NX=M#*Ӵ:|9RL5&vkCBpm"S޻v/yΨݚbPEtQGѥ =Ёh{P"ٟ`AY#$NP8"?+8^ 'q<;W*m_x1bAltNEYFTtZtl8<]8!P\w t,:9ڷ=rDۣqäpގGډGsn0CZ=DESvb۳@fCTdE ?u#8Totlc!8!6F6FP,htE'O9ۚEq8aJb cy4a*RECmCPW䀆V4.Cn(tpEK[K?(,:dkqV+ZbLٍ,wP*f236|XLV|ui>ʼnV!]+@5lVq˛mN|~1?x{2O+ 0$Io9 0Og_~~l2Ks_׿o~>^ϧ+5.33?u2sWs7 ZUT/^/={sh!P3sv4­,O0ܟ5f ~u 3|KסrՏU~/\o%ب'm3-g7!p<_.?^<ه ժ6q%˜hMT~ Esǹz6xT|2>ھ,'#CD $OL!BL+y $0_\ %{^/ #h w*^(t >ûv\o=W8sͅ6pEGi etVw4֒r1ߎ7njCykDzdQ l}z8gFĢh i[Jh+ݬ+:7TehN m+G)sO!SƘM=5*BM Dh[PMz^^jJ%L("4wTKk4g83ZjBTj9WkI>7|H:4q&PF$ؤ>sX S,B;l@30 n#iTiCJ1c>c4%NT1%^CdRoR8Dic RRц sێ 'Vr-+krHO΀]=lHZcgleeP 5&>'  Yvw&8ACEe2t0<#)xD4M G壖GJ {dH 6:"Ej8LH#So k.38TSI"‡I*X'Nro3'5Ph DI-WrTc#D ^YW FR ͎{oA <߸lu"c򷐭DQ"f[HZ2ՇTjw@8P/Drr/zu p9r b* [hfzw]dC0WE+_Oxټe0Z&hVH 7LQ/I\dq.LJqjɦTs3p2 b!2=5_p&uhb)X2agOӂsSؘfͅ5q4Mo*YhnOV14[U]+EG8ͷzغl:N 6Ƌ׹֒h,8plP4kło.?,fhFѝ2lƩ8fT@ï&+sb?sNn:åbvO 34b2۷*%Ws o@xmB1ZRM}H5 '!]ģp$9dZyK>ݝ=<7GÕi10T^ <tN½0(ks Vs[N̾n^2:!^ˬ.z~&Yt7G=tJA,uy{*'QUe~b$uQfqs'Y33ߞ pN@3YvޝEsĒ`*T\_eS*=*D$[YH.8wl\ͪ+`}/O+'jTF*4~݅yئ]S3>yQ D0*{DO?ɷiALIjVgÌx)Ģ}yv?Ä 4";N&ӻq ^A}d8l%dRqo\^eKCs<}|hN$QՕ`DV!(:#;į5£.g¶_#S{1XS%d {?˗Ixʌ2#),'Ҕ.OY2YRr͔w87=ꎮXg<ރs ]I /S:mcT9Sr ۷j`scS6GO MŬFBtWr!YXwk߂L${hA қt I/\N$Y13>@I)]^m`Q/D>wD>=s"״x]T8\hҷ{ `OFWX)Z@ qcPM]*NBg,$mk!<,ɽ#.u4jN%Zcm6[:j"Fk(lse#f&PCo#%Aw/˘N&WEIy'OIc}`w9gKa4diUU- vyrZ 0Lj%Ǭd-o͈f⣟Ϲ8qJrGcr2G)+YnmrII'T]ejKxbvw݊6g6s}vX)ыE2+>CQA1=$ԴLטx2Iރ]5 =[hߖQ7ӇEG ss&jOK؆af<.ibMh)TzR+!3/-(zj)MksZǭ)ދOr zCa~mjPF,zrHa}jNe t>G\.R)*YrhN;5y;*sEqԧ?)e,\i V؃RyF->a&`VW{Lw'Zh>S̘6{{UB?sB{M:JjG ò#fx{S S%U~=;Ƴ;;'=-rݬ-?ݮޖ'VS0:nz^f~sܰgaH#_W'AZ&D܊u& o>G}/G3ͺ% ʶH>~|ƴ 0XCޛIpl AD[l/?3?cxu#103z?%acu|?4>N0qjeB`nu55 Y5#j , <Bp e|3Q3`T pZ Zg;=LO0( G .v`<рEq~(h'; oqʹ2}- 1;s}g,mQ3T6|IŢgWǀ߿&(=(nL\9 Cʴ=dpeS5ū+xMoFŧ U PsBj1J(! O:k#Bșdl4 (>J Uo(#Uf{=3/݃֗ٶ\6G> utZ\`C&*YZT?aZu񴀻8, ;g…ߙY iLըS\: :2Y.eINfG9J2O)?[!hQQq~l#-s4o8nL!0f'6\ ~|Υ}뀛h!34c^i9OI94M)sY&:ih_9bFfYe 9t HZ$P~{U}"I@}]+2-dcc+-E 7ljEˍ9#+KYlyD9ɜYul[#S·8R:Ad!lӈ4al)y)aL L<5^CjeYxvWT+t "7p3X Dw7O.ޚʁSy_.^{Khۊj%v2-y<&K~L8~13~Nae3S9MSE5Qn3hCfՈ1!\F=fJ7+t©\`VJsu B@oq)^e:_K7Z{[LP@qȗyYy[\NNjCOoqqe%Wg_ })Z ѿFiem-#_Lzcmi\T^jV~?~NqSS(A?fxln*뿛8vv:o9o*mN:ҀႤ,z D˹YI4~4008 Cj39pϖ3ݚ|瑱Oe\u <ڷ"xM.޵q+"yOWb(ޓ&H/-%7VX$mCfqKdAk[.gpf8g H6e^t0Q SRgo<.i(fn WSyn BfzfyӒ\Ǔ`E|0kL@n?gzxY,~3śп )DɆ>?e ʢQG y"Z_$b򴷣 ѝ;on0ߨPoGn9-h*mA碸eq=ѱՁt>B8ͽ\0r[ 懒GW ^7Dq#hT zEjSAҥ``m~7>O6nW/ XXT,P_6 ,F9nnTzY ~xo6xa~X.Vg8CF2":ˢd:ȔP37#/ @t8ytg L$z0N8JCBEaCzƙg'!ϭ7cަ\)#Kr@ж.M?v-"hн?8zea>w\.~?f?ų 1T,RE@%HeKm=Jg.oK,rc~S7rٯjZrC;NIq,)$ō5X vo'cv(XZYOaLg$l4Y=f%L90>k @Js^:an\DO1$7~d/Б8RZ.Rg5CD<!!8P c1c @ e$!fdBdb8 8>2AG )œ*c^: ؀}}?ƾ/'LL%4DLIP'10; VЈ d**(0”;E1+Fi^8N ͓-1 _oHY Ys,GsMy=9jfQRJ|s^kzUDN`&F.Dp֑?]s%4Ѧ7qsTm9VŭBbF1j%8d\Wɷ /}I0۞^ؑx$.wo9ApcoPt`0_) ?ySC)}t$OgEnenX:%ej# Ϭ72Y-PTݛy,KYyx؜:nS|%h>nGu8A(gL}Fvt1+4@:VID(& ̣D,g$,踿U,FR߈clV2@ƀ,F=4SY aQBA(HԸ lUAmpp=\]&d_pcz4s(NHH@",8 %pwILjJyL(< 6^S.DffDi{Ai&IF:efyI^LAne=MX; *;Z.8|cX嫥`%"ek_8q8}}f9Ђkm_s{ZlGn*P4ՁN %(Zx:xo TN}H:H"tS}D4u6< GL Xa._* nwMJLEuE`>ό ֪jA6'nxGYN)=C8y܅sȻ[YKpNitE>H مSW=vTӧA TCE } !DVu'꿽,3 8L*J2E3:!["fNJNbU&z2X~I$Pk-g "hB2HKy+kމG/w)2+RvҢ8F b!LalK<&qFR qbDTHDޫΛY;t)p;4bZ2ױKY y"%Sn[xYc!u*x6zN E]/4!Q1¤'T;*8)k'?.ea n5ac,GcqƌnggT2a/3Ue7۳VY1`WT.MAl4MPJ-aF,ķij2޽({A &1 GNo6 dRdȈE#˗%SQҩO҆  A15r".G^Ta[{P\ 6mi*(- GGS|电6rd=9 !1&g$=mJ^ϔM9OgOR"+ ,g*}#g H@b\pXm8`UЭ1lc'$x-q>c)lʹp#JI$O"N2#拄 ť ./Qm΂[l]_bh{J.b^/3ȹH%*"4fL2N(1N,3g-[u{2``dJqJR-'H%bĶ%M7U8tթ8y& Vo:v wdz3c]:xZi ֚/iϓXڑppבx1R(9C k<NUbvv<{^ȣ;svB8ZSVJU 4 U|[2"sȵupvo;3@[[0qw>ކySj@QR: نCұ'{s@t7@auvЫth . u>hjP F8T"^7|zSĜT:e-T"/l뉎m}`:?[Vj\,^] />8XĩIhݮ?GO11 yVL2v]%lOv/ӻU9޿nlK0wKh1zk<̍>J&q;Э݁nt[ށVHiujK,(SXH `Z4!R S CT!q$E #1VR>}9źR\X/ڑm6k"73p̱L5FZP[єH@ԬԆ18I+,6W1WxL_Nc X7,UaavT%B;vԦĸ8.TDQUڠbP*3e*ND2N@&N20`%1 y|̌AYe4Z C-%*LLb)5|I@Fm6+-uٝ6_kaR"Qi辊v΃#e1p|^=B<[:a{;4r9[\^\@C%#0bR/t, (aDZѕjb@Sl&w1>jy0}54(72W$^,]u?rtwfmބ\{_$앆aʾ=Xf*cp&*+`@L!;,`#`.>B|=rq|Iȅ :4zXbN w hñxrl]icQEH mhm?v+|4 Ok=Gkərن!oU` Jo׵ mNPPaP w/8B1#i "گȾSL'6z Ƌ$ZϿQEO^J(WJ"oL LXZ0 { ``N4Ch[w&J@)j=c"@PUzTS$;Vώ.q{tً>L-gfj|(Ǜey}%}TG{wzZث=M/KqcJGOe>Vn4lnULenE[1 $YDXMI#%(1x֖V>yIIN VtA[ 5Rg' qb5V4#%\@20F?xjK!Bqoz^L- d0n9|{G~3(XOH?J{ҺA#ڇjNHV5-ZN{Cۖڶ3uvf =YpUZw{B6 R󨃘:kEלuO 1udaV0JI{K,FitmԬRi ?/ALys}IѐѐJ&[*> wMe^!L5I?q )RW!$_J[fJ9pT寛DW8o~&'K2iBuقJ{I`:nsˀ]Fvߨ CJ w8%%+L}N񐀽N):lIvi:C-SӜ$hh̯DAI5I B28!hNr"X(q <58+SlU0gm*W d?k\/,jhJ F!kTcp2"5*ZuRMaϬ1B_i*d.h؊R!hX(=&>xjL-]?!oQ[R_(i {I̚'9~5-$]mǫl>4X՜ɄϯYZ²ZV[ @Ej~O5d=⟽y`["8Pʍ,fjx9_LvqJ{Xۇx LIsĻ<̑x6B=LũaJ RlrOĎAPD#PA F|u3ta&~\ㅙE}A2sv˾/X#@^ڡs/80E{p\r-ȨSKUH+ TəA)4{" 8ڂ8Әƙ6z¨T`<a%1t{?z 26~8lA apR;2҃g55.hʋ:YoUcqEV?/QS,$ q]otߛV2֙iGp4JѺZޙW4$%[]>7-"=?F<;h`ꩦ`TS?Z)(t$`M߫$ArI:q7\Bx<b H45]JĶo& $t}-f`V^Oj/K͘ƌURs!ؾ Ji)l_CD{cj>G}C~.ZO F}!@pPs`2xQڲ }/,EexxmdSiE}}W J}樗I^_>\>}.uЋ.^ŷ*N\^aZk?BIɏ"MpNGCӒ&8%Bs22$Eh\Kiz(`]JZit4ʞ."Dq@ifNz,򌗆Z'+NZK7-dD3Q d*^ZxifM8p$3wzKEo'L C13M\`f1JeC!6 ~ [2-dg`BVhZ@ :<4GJ x72|JK.JKvȷ|ļD8d* %S*07L ۨ} qIzX4^g{ ߟ;|X0?Ubv!/8+$%~TK%*JCZ”: JZ#Rm=>ڥ!hm w00muAJ3۸jG#}8l8~O YbhQJkV(t\r50gM҃`D`l76ĊtQ&n؇([pb`":MLjZt¤:ADTy HmbUC]x%SI PȈ=ⴣ}XB/\pgLJ]S]*O[ԴZcE6lϓn\\8X^E|amia lFd9\3Gt}U1ݞ7dIXU;jv=_=U$ۜz-#ǢP{ `P ~oyܝh2cWwݵ //cqtUŗ{쇧"Bg4 I> r>{-Vrp3?fK/[# Ep6+gP.q5Zuf \vfWj5 /?i!0=.{eۇFFдwCZg_Cw8'L4eJ!$`,*03%9<歰떏Ն5#hjc2LGn:I$N,Pz&IJ{=( 95sk 1߰tcN$ڳͤQx.4LBZ8!4tHtĸYQ1rY^F䔈s[3^ԋofr`2aIAL;)2esP7N a9sJ̩"; J#rCXhV$($R,I D47`&` )ʘ556uQ5jj5][5*I;콯߾[arg?h|r^TCk:Q>5FάJe}zu}Q6>ƨM.ew81+1Ѵpy4ŗLu|*ZJ象 r͟i 8@ Ou`H0K6V8w0/ViQ8C[mn08۩u,4ֵn:oCQYܢ ɅR_?խlF༡Se8oN'||tT}\Aw۝GR*4\qw(x#rp~ /"'#X9A˜hrkp*+/{뜺,(fR2ԗ>Y{;UW%0RuMX@eP2}Y6?H*y2-oB BwVUbGG_? (BiEED~ϻ, E3h7״N)48^=]f!4z4a4yw&>7fD IE{*;I{ UYVZs"̻6 % %X1&\4q4c%uM8/1j8rRbmqWz'T=rq8Bxd=׿t'V GR|D*+hz=|6ad<'B*0ymŹd5TaDM4y -*-+&A@2HO2/3cʺn .Y%2x0>mr"C\OsgBqLJ+;Kv c>y2,6}n'omWV v\K%9 1O9G5˹S\fU$p a8љdW]] ޞҔʙRr)喩:?X0?@(fvUwŪܙ  SvUl ضZ^*d;ιUy.r{fZ#P.=kI'Qnm_[;%p-V$M8N9Nӌ*i* ,s -Y" 29YQRlN0iMor=@'x͌ ޙ`ytȻ)¿;H jW8[0@{Z%s x͒ /4V92!'UR%DjK5`7ۃG\ZcȘm7XXa_Y$R%c0 p2KY2Jf.D,1ИB*WzR_n]_jIUU|Rܗ z^;.{pqMxqb""՟C'nY^ݯ`++ X{iӧkqh<(>޻W~YF$Άdlk7vz0VpºrW=OkO+ɦlR>k!j]vҊJ#kP0fa *Fqߥ[^__r/gb?_/W%g?͖uSIcLzaj$WF쌜2݆+g䔱L޴`#r_ 6Qf?ƨM.uFX]1l6i3NqQ3׭eSGhdNt]ԐǟrDƝn9">y\V([S>Ywg0VJC8v:{p!C!Z]ϥ QAFۮ'qzn!c՜nymOxyZ׉*G>>;Gdx c $RJm_&C›r.Ċ=E=u uqo?r:y'G4әmfjP)z>9 H4 4 &(Ч8&h9q[91o8cqӐ^R^g)륻4, Y4"6Z|k&*=y*(akz} /7aqa'>pg(RK97R@.vDsw3{5_Znf=jCV%A[AQFWNyuV'HP$jPֺ*ҏҙgbn#H_XK:HT@}gB\NoF.<:a: e!~l\'g:9ٸΝe n2OD{Al4*Hd2y\G(526I_iֹJ^:nB]@r:"X%`I@WD3|ENY2c;ňP`19F `KD!T@8fQ`R6j8t]/9Tn-Xi+ނ5݂y4%W𧕛8(}h $G`JJΏvwւLk%R1,wg=-m |iEֻҜNЬ PrTL`s‚xX 79htGl*rJIwʊ\JZ)䒀/ȓAs6.+ ĩF!>bMXTS0}YѧosBo:.% ;%& 84BGRMB!;@F _y/t HB l-(tniS :QQ5ms 5|fc5XiB j~q*P3$o>ll<W,2k?37ΡdC JgJzMpR88nzS4iOuE-b@AE\ uڈ/YYR>.?'E`&0D+mPbƃMCwǴ!_ 'v2{w xiEdBwpZ\* IkR"[JgBc3g<Ԉ4EzK]"(X4Fb)61'\AT2K,gAyżg͟BI f@%piրz"S47y91C,|2O*qO4fxs n,T6Q DGq:`L4^FH1 k"tw1X* >1 avHAJI $]΂N^ %hWD!` AJ/eK|RR' u&ӢM$F_iy2MXÇj7Yazcy" h! @t"釖:dB߾"E¢⬖RU~ǔXPp` x,XM# !+,EJ HXiALN@%*"Ĩ>.+Q1j Q* a#ʵCZT"c/d/z_:-ï,w&)rswhZ/ #b.S ߗ3}Dm<|6A27<2ko>EVtʜjVhaj Gq]őtho*U[}iFi@5B<лմE% 25A@qW`F@P]P_zrbG~8~: ?ObVF_.<?g??/O Xôyhy}!QHESMvz2+TCꣷSiݽ/]f%Ǖnu[☮l|;~Y(b18oɪ7vIH4QĈB]1rPy]Q̈+b"K,i{IBԮV]UBUvQ31*;& D"s "j8V;p[{<. a8㲠-SYT(`]Ս; ALK'6rnB݂tׂUIv'Ugb[yIXkv~ 8s.?NloRFɈ~VS+̅mTXL>} z*wlƣ!|US6UZkm7Pe|Pq^Qn'w$ 5ja)8zSj7k5o C&oN~=I%wϐxveL^~o F] h|{7 {x܂UmD^[/__໛;E+_O,c&ܾY?6)Utǔmlxxn@ci: nwҷX쭽)%;pttI)p&nT uHl/vfƽBA~&alWȅ-e(!h5_Q0eUFSBb(ZJkꑈښA#\ e25 &}PYKoB{>TtG;v>t_7;ZO_hT%/DFQU)u}EW $%mi B Z,( 4BX5^D8)i]8P[^J N:?:ۻ̨rtrzo()8s~050O_>T`4@ F;RYd\$(*XxHSJ9sQAGUV 'qdx^'ɉΑcIHFཏ~tI2Eˌk]1ȈĴF[;*'ɾCJ+na8IVf1i&V$e똽ֶ'˾urdb%VvT=ENnqmWbC'Q.zMVuk}' o_go?LT"g;N#Zy%ry>q=p-x͠-TM䱜 Yj?leR~;M)~=$PFj&ᜀU xQSM}׭ϛ{ 4F7ݫ/$[ǟW#WFQϟW+{_RT#yw QsSP8``us:="I|˖3i쾹}? s\bs;.i$wflބUcooo %#^j'ø`TfltF@{{\XMΌ6wc9H`v~8|܎C$p,rsF4ho&MpG7`3_^\xB8^w5;ȐA@Z$Rl"Z|\B%e=rsj(}y/`\fdj#\ {ׯ^V; `Ma"!#ykTjFu3iS0XcEJaDx1f_/Q Mb̯%Ce)\4?_xQ /.~>_jxef\E }reySlEx "$a@7`UAb2xgZR ìBuv ǛUf\e^-T9DO ^MZ`|B#fH3J2ĊTajŊ(|FxP|#QS#| crXD]A\]AyIT@hr^c 2JYP^1oΤx)2*bS`[w^PJ"ARCqR)0aOdmd!-Cx9 "2(wsC Fy&4@kMlgj.(Ev cN͢NM%9ŊģObZ'_J.O#q2ďa A+QKI QQM!~ٝq Ǘ$Fo4cAyiYV<GLk9'g=PF/[`fP Aޝ=~7cyzPJPTD(gFcUb`ݚtu_ 2%̺5h]B9Da׭c[SԁN16n˟$r3ּRӺgFsrMV{nMyP:Mĺ/ 2US[К֭ Su5ܪ&(;e1s՘9aWeWc~5fN5=31w5v5A0z|5f4j]UMF1 NpWcjmjv[,]1 1KByWcjmj$\_YjƺsWcnUWcN1sWcjmjCWcVȮ՘[՘sP}bԘ5f1w565_Y3՘s}>kZpMjϢƬWc֚vs@vki»sWcnUL1S՘}kS՘s@>3|'u5PcLqzt5f%]1 Np~fN`zl<\兙w@cC q`5(t!!jS\N>p`OCMf'up{e՛`Z4O"*/G<(D[!H1ě sJ^{MR&,? DA(I5282"B>IAj2xC caGDDTBέB!,ӶkQf(Ø! )>~j"?#!фD$s- eXGaY҆6e>~4CTU)A:4*-%E<@ ]Pi9QGB&&!\U;+``*h=X><ȳH^ IM>므B-vD(6/4&~)H`p Bʓ뎖k! Q X6&sP\,,"3ڜ~C,DX(GVYtl÷7 lf29m&c-p".0FDxzn'ࣥKSa侜S>2+YMO[˫3?b0mczۜɒ4^"W{ f-^G -ouٻ_^8Ί|3|'C~ghf`vY;%5D=XR7ήι3aМOpk| ~hM-fd"(@!qnx)C+œ[Y5D2 ~FFwO=40*²hoJ@5f fJ}J_͇^z15du) 3f4ܤ#ܰrR#?U prr?"܂#[M~~/~RmfMWDjzh" hMi_R**$H6LgI!Xa_,vB8#:?`MPƎBW חZ1`gakԂ;/{#i f̪UzL}d5\D=|iYNi8ȽbI%Uz71.TsJ͔ic\̧!ϋɼ8 0JE19=9p>UR`eJfdw(?`QҶ9kd~^to]?4AẁuhϲͭMR#O G"OLrB$12٤;/W,LoYE}!RGsT @֎*߽| tSɠrvwPnD1&u* `q Q4,74E*NJ[lXJ )؇QSJNi^ cq63#oǟt(Kڒ*ciJEL6 (ܣ0Mu%)@y^pK,$_i2O/'?qmpSХ)c|+rр:~"!/W.]d{Zn1~yu$)QxzW`ۖ~Loa]<0/i="IZ(`\JoQ+p ^7FC2K{a&I_r?-""Ep/(Ȧ9ߖjא%m?,qT˄N 0,v].Alw-n0cqV 5 D_Oщa<=V-$5y"Iv{HrƕLڕsON`xox#0ZM f)MT$NzjYzjK *hY*y0ecICK0?GHmeVn+32>+OREM|Y|ϫ+OQV]YKsφK0S?2'e8OpἪ `[+T&՚Hɴ)x1OcKDoi6oKAfRts F$t]l a/Yݒo}eXݛ1c(}٥RiOaMIŖ/0s+Ef&qip΋$uAD dqI;Kp)3)5.;?cP#6-zn-s_R1\ߛ|7:ty i%-6˟3ɹ&>m=qBJ%nTnήDЯ# E>5miG*'7cr\ +Ow}j`V^sy004s v!gÌbCd>6:4S: Wśwoh6\uk/iyьG}~޻J+vtKz{*)gL~MS_ Â7t-Iֈ݈FKEf8e5ܙ6 c'&t^R*?/E(&8Ǒ/kmBG!9rc\gqշJ xvBQ?n}SYfG'#alm ')OgkN҇ԟ$ mf^<<< j?8, LE2H!'Pa0!TgF"!Zlcj6Q?mWСj"ZǕe9:֫֩ F5E1#R*ƈQ[HBFKZM$r&Q"bF.PʏT+ ,t:1BAjh[J9wFH) d;XP8nO5GVX1Zb>DW(2؂A wl9ܕ:s; 2.-? sH95O\D. U|0f~;]v {;@1Y kYRD)zvO Z !ySSı"k`\qrN?fzFVUb.w>8AYuf.MWmv6ΑF| gT'`/:˄wyFrtOӂy4$Z?%޻ECx&e%~k9&b8?/ ef%NʇE@XsP/:0ĥwﻚ1zH* :JIF/gA\-y=0W'8IaJ"(RasCNh$`U=7TWgZk!1Cx=!r<+t^2HnOȇhEy]6WTrپTv7%WS zfwf4Nנ^DY@JWP пFv3KC´E* M{j̲Y0-5kC_響x́wbC4цj`#ݝށ+ٞ^ Xy$Vr'x6iV_ǖQyI0IUw\/*dю Е|_>pqo4(Uϧӧy9#2@g FdCk 1*?=pfߌ)o{jBZԝ Ւ_.9*hVEKbvŏ{R!,ΤkQbHkz4'cƈ)pq@q[ޖ.o]%'1Jy[s$.n2dXnmK. Gs/LV;.bG4Bg;d/w@ !1=$_8`_7(9/ܷ;h+R(u+,Խjw(,z"ΟzT~O5fK,j!ܤ*2ێ*`!SX<aÊB`FDO%n@6Xc޷oKd^6"[-)Mi[ڗYV0.-'DAw]O>TCvhgKp$Tj3UP]Mͬ1]8Ƹ wzQ嬠19 QI$$ k+Txs xʉa.G5R_T3+'35\1wQ߀j2{Aw[;Q LvOS7BV- 轷8(x+E7<&RSdm>@4BDW4? :O&sLQYv+&'D&zBI\Yw V]V@5SpI*DJ:or$\aS' OE.8u(4NNL}QRgN» )!feXCŅNhzTm72?bCKB8D(/ c 0.`@Ls*B6"VѽF((|ELv)$`n_&_L H"U;14XM)c^6_T$ Jr:*)L-sh n- SRƐ@'ԡ6Pv5p)|A5[tX>.+rf++Wo[)TS=w"[$0 b3U@Zf-IcP@^R\A+)PDC\OD9s P@I`^TIW{D#)E\AQIr0#C"NkBI*S--F叱 ̹kK=W.R;8k@:`XV)͸*$# G\05meXcCE.XrD9ꭰTzlGiT^g 8ʥq#_05G^ڜ*,<ʝR) _D&/&Jv`K: zKaժY^kj0=Ip-\RFBj*Lhc6CQ 17Da{=D$UV&ajwt)-s?Cy22#s?sA`ySg11bET;opQg%B_擗[?2'onn0#7q~=NO 3~vG5 bge"m E͆P=ȫ{>M9߽F`Zp$2&Δ%$E*@2, Bk$vdG@b$a)qJ׾U.(6Y"#Haz;,Hgl!Y]0S-,Ty;%N3Ba5ć`0a!BɐB,$wЌ2퀡T>2F3io)k|n+6l9s,4C\ ]:SSܱGGJ$dM5)¡Wd/ Mc :rZ| $¬C'6S +AaW^F2Q_L\F9ApFCrmƖCM JmxJT/@';S2*C ?;2$:7ƤN)`JbrH!*U甇GUN#Pߘ(Bi]{|ȋ8G$e" y%dflE4 ϯ冶c#Jhƴ8]B uX̬&@VsQlPAkgzYQ}z@"H4>)z PT4o{+֌ /^dQzG{Hd )PNrĄ1i6jҖغBU|v4ۻH"ez}+>vg ;3dD$*T"s[c# -5(x214Sp54 clK Yyֺg=qYH9ťђ8ɿms +0HY: Ȁ E(g0KWP#hURLO~qÛyyg.mC{0mPŋv`3&W>4Ci2MμeUFgƪ<Yl1P")+b^^{{*v ƕin͚ ^:o5woCOҼv&>+!z2(ΚtsoOExW{y{^%-#3sj XK^G )b #4 Jɋa$8'k<Ǭ93e8oU.r[Ťtsw8/ Am怴vnHK&J "`JeQ"TpT#;b8xP3sChQPZ@)GȬD&FR(v$6/ rN + 4\8m^0~9o(ܽQ;\\Mʸ)ӚL9c3r)/kHr{ TY9 #31o 1c[Hl'c~:t)}aU3HwaFHyyXWP^_ 41G3ow⻘ͻ7hoԔejaiU^'~=֕s]þ}1H13Dŕ{\$j &P;@?HM융6i׻MwGN1QtL$5ӝ /t}쎱Oi=ESWI\qwϓ:pG;Zru$A!(<oƿ>ן&߿t&.QKG =h?P^Z,< 1Tmw~m$A*ʻ/۷W%B+rj=gq @QºF Oi|5UZFT0)F4տ9jl=|1fkrA Ȯ(EYŀVI {u8`Yn/R/BcW=·_`/idК%.lwaUJiJK*tILx2].,.gwIBe29.S(-KE;ԔVnV;3#BA{̌b$HD]ˆ JS+[^&ӣ ڬw홉4#ҡdҗWINL Wҳ֛$@Zmm( tL j ːwIoL8(𤓘"^`R am^4ꉊxiT(6YxUa cSެushv.2x  v8E Ru<3 ?ˣ3ny#IA(ݻ;]Z~Âބʏ@iWj#zfw.HEi_b=ź_%|,z:O5)R(DJ/șׄ[8ڳRwE>zvd0N|pMdTf_-,߽P-b,*K ֋B*uq׋FN!9&3w~* Zi6AR <$tx^& I:_V8:+f!FKY1M;$A_pDhBD:qTå iWg 8 朮 )- BP&2xx_&3ۉ&.iq+u##-\ Փ?y|̳٧?ݬ~\] ael&5`6/25-8&K+h}+om>*5TX^6I@p6dY?m/nXÇ',gz _K?y_H|uo8k&%_D_lbOU(g5(l@({ ɫBm RDSڮ-s4b&۝WW3X#;MBJ9>t<1W.Y[mm&,JeKM~HP[)χu1M悐fy}ElBR"Zu/]I`UG& ,fDg81f] %2sGЫSHZ9oyyFs 1Mf'!(\+EWRWmqB Rڋ"k&mmwmH_a]j x4*؍}[[WIm˗RaPg+54|I;u_f%Ay5Q0]En-; &WJ O &'m5/8Z.(*_fR-DT1Y:HC'q,^uhTh:`z#tҽ7@@MYՈj7o-pXL0v2lN[x3vZGmǒ})-1Yq0g}E+4>/\ۼ򝃹}m5 퀜(22eB|O\KƟ?<z_lCV>OH?$AkYyml|I࠭BW%mzp\(Lұb 攢>d|0qB[+hFJBr,e߂8"QDVVdQ֡AL"룩Tłgr>2<ƝqaC$d!rB/6Rs -/.Rd2õQrc3Q&h4,N*!J d}}/_lnn^bB``+akq<|EEz!ͫU->= wsC#_/ $KPտ7ˏzj 7H_[M/#}0c ?1G(Mg7 ”s=|'$e(8]e-G1) 6 cZ1z]]PyZۆ,FFlPӰ2kw]>-cJJuԅB'P~FVe)ccIp'0(*#ڀJ^>jNBてRߊXMAX/W@F]G2m5oy-m5 [yD@+pc4IsFRbK+NBl*_J6]yj]BAE'V{GO$W}vg#,SO@/u!wtweO&v7Wտ.9fV͇HK|Љ2A1V*VJ^*.K&Ubo.=l ] ȬSa-_/ CmU*UwFhv}@6L\^6 ]Z!AM] Di+'qaG,B]09u͞p2- xrs1N,$'}ҏJZuu~<149Uԧ9[ iS}ܕp;g/Od ?QploG=+ŵ qv`=Q=(ag ?@#k2ݏ|2+8-tyHG9?GP(σR>ˁOcmX+z_z 0+w6;+Oƀgg'&2o801mF0?^TWXJ!%c3Ñ5 rwÐ&1 2]hLsԓY'] +@7!0!ϰɌTCah? ZB~X"mqF >ΐVL`ƺ1%F>%?4S9~ ^FhMN\]Fe6W=vi_dHV5I˒~аW>ī0zwM! ^T{95݌Xq=L7f?%GR>xs3.YL/FiHWBK-L4j4H1vydv6mSҥxjV_v&.͉ @cUX8Da hcS!3Swx_0Fɗ͘ҕ1/~UXc=sc9t!#Z v,<j^@{2l]@ku9VI&1hC2#,8\\p؀Z]Yz}8J}DP 4pRn|mS8e 䰓ֆ1cÎN%?dw :HLg.X߮#-8`#۪ QYw!U$@ k[e<ق~1rv Z3^r#1/H.bH!3Җ}i cI%2ɍׂo&@+ć:H IWxv.P8e[vZVr,Z%&$`y zgּx~5wnFiZ#Q}[wn:#F~$r3Gc㡿Loߏk{C{i1-B8ی2BgoT-F݌U]ĝodudyfyfyfyMubHsB[z% I#dQڀZhIUbuAqׇI(gQJ#qCI iڠZ*bDf]fV?HD-e9D+c!%QW1T`/ITT+v}t o؂ndѽ6ڏ¶PSH~Gw̖Ɏ.F4-)k~s8f@nz9[ -,I3NQ\I T9ɘ!Aڑ6g<ʐ^6TE+Č ]y!%'Hu&MRHB)CZaʡ^Yݫz}oA)+5~oS*Df%'Qd2N[ƴ^[(0v egn:Y8:^<.ibͻy] %WSA>%}=Z։JU3O*Y#IH::׬ (bGLWLHW~ *bDᘠMH)'8g^ie*29&DoQI*[ÿL,(&As>`>bׇp|$X 3jRsYH|puUxC y>MiiO($﹦؎Yr*ͼM f:(5"w:DQZ Z ݺ*Qf7Ю&蓔\Y`my<-NQ+L"D+ ̰hj"7(K S<h*hLdTN:g3q<)29c'RxLkS)VzZlYe%t$qL$#J@T >kYfj6haɂ39ۃkj5#7OMa7kJPoPP&`jKU. >j/z@3xvۘ(B҅bs-Nk,a9]=E9R0A^kK.bEMK:/}\ŀ]V[mPa&IBD&'{P䪉Q a7&~f[2g͗W!:zˍ^=ޒ^\c>X -}ya|z7.>3CDOmZ>.m 7+5 !B1N0ļ5TkuSћ~^]ZT{7.YŸxERǖk5Ndu:v1}sqnz-h>_}rB>xT0Z qr}UW͵gy=yG?)1:J5n>wI?&Y`~8&en]j#lƐmbmɕY|tM]-6Mj2vഝ>t/N3=!wq~[0\cfC2@+D)*̮C8u2Òx͞8M9Y!皋Wj|\ Lqhnzs2cV?xGӮûfz1ًI YƋK,3!sJgcJd`3m%Ì|'?u6i` [?f_F౤b>GIc}zAK\/.U%2b`0i?jgsw'aS4ϳt.J[7oևIx@0~KdoQYJ쩒%iG)*H H>zW->˳y75[\d^tIJU Ɂ+!:%@Uxf[ fv;ͨVQP T?<+9ॐJbc*ubZ+kD09SFI%1)$՗y_v[wc놉5NQW]8] [q^p9@?orlt9 wA^_۬0_aeCF kٔ8S8\Wқ/`(Uv>F$$/^YOh 1m Ty1bm`v**v+(g6{v9[PQ㰏p{?cz'jUz8j)u`byM6PI/eu%OJxPIšn]Eiɟiuf9 7 w\o;Y9GbWI*ҙ89]1ip& 8B*ʆNKkXns7cb# *若>O͒Tb^ gi% FrDcpp6qfŒ2笵XCBjRFμ5\X)lC5RrabJGM)b5Dj DC^.]I%X6 gіҊeQ>JFSiuQ< >p"N-Z7N KXK-5'iUah *{yHE"W }B%Vjw(P%]2YbUe~U"wUXV)'KQt"Ie+._gY>(Limv|-F~QCgŵ_*'*t jV©lUA|CGܺY|֩q=h~_fl+Q\xgF9zh2#Qtiu6J[x18L<Ai { C+zt/ƼLqʿ_^y, pYhp/7gpxKvܣiqY)r{bcC7Rrf'oP/lR8_`#Gj4F馱Jc5vg$p<4rD[}LeG*/>^whӪs6Q*4H^]ϋ@P.S{#{'fLoi [M~7-]1̸NA}&l .5!G\G0LH`.`PDo54d;1'@'g4xG:fcA7*ax뫱@ 1"AF@±Q6`ܫ1N͘>~&J"#4) cHLI ~ XFm-} D۫m|WM\ipg+WA*d+dSWAMqmOgqQ*dMpKWm؀;%CO.xUd⯤e*kW ThR`qq@dڛFWqG-ue.:嚗dC\GA6$2IpQ%KtOϯ{S&i>|!x NSu@R[UkivW30=,7n 5YlkBǕUwG˚^^WE?H`J״qz@8h%Ӎ)M5fҍY3,HrEZ7dD=URbVt3q.fo 0g4Y.$,6 'tF8_,eܲ4*yEr)<ͣ%@J1w\*h݀cPʃTL7bj/* e9)fq>%c/nK3J BB;^nRx(B&U*3&KXy}`#0Z%K:q„5c|RX\CTBHۮ׬V봃ʒ/م?ir.q`m.'s9'k$*ê  t}KV4u& ɐU8\zE60nJƵY_Ĕ^Wq׍Ѱ@ռ.GB}!IN.3nu,&6|a% `4B'A{Xb(kQjnm̘\J8\p-2$Is%Нޜ՗X,}]UDKBH4 dluh-] o\,C ;*]rthQO%ݩGX08s" O1i9QmYrx;k|\:GcL}n#uu3cg[. XITKOk5^+s+s R{h{@:fLan2@XWLjdBwWDl({>]>Lp Kiܵ"l--jqX F;vAiԞ gh@ÎcfSiN% N6jǨՍ7Lssi;Ӣ FՆ],3G-|#d lCڣFWn|yɧ2_C|#&xB &ӔSf /9<rN1J=1-Fe"-DZ˹TP 11H$R"u&% <65H yՆ*yƗsR?*)R'۬JA9 Qa g]3ƕ ͓gUB]X+$c~:d9T%o]+@!,#)N[aR04c& K vVk\$,x~2b)R*Y UmRbϰ@|?=P[ 6'F?pKZ =@`87,0?OЀ)eJg&̣4"S.->+6TmWl|i1]%ēw}CV*O.Je=@K;ZڄY)uwj8+rJ&÷s^|uWRHV-Խ2_g[>6ispI!)MJ,P0364H$)ǖS!LSp@Z1}]wl+ \I\JW_*ҽUto܎P Y*|tg1ILb fc0| f!P1jREj l~$׋֨O~$_víR2R|`aIƉM4@ƄT_LJ=U<72X XR&@?;@Hbw6R| "+xKyyf =Q" / ` /"*X PD3`8ԁ(aFi""fAahh xܶ9݃=I܃=I~~kB3 *務-? 9:YUL"*dŽ\aĥ;~٢Vtxu|v=jŷ/I*N sڬ4 1u"G 8ET!kF=4O_e5?^0_ãW{'G;/0Мeлh>0qvi?ia=;> 㝣?`uvݑ7.:| ou&znJgLp5;7*;1隌@) 刘ܕxޛ{Nӂ>?ܝa{7{?zxt^AC/NZ/O٧= p|оhn;hGc׿l6cP#g_4vk_p\+y5*:_?fp/T[sm;Z ̯x18hfmd'.s rk=@Us4 ~}5u \t >]G;O.mv^ߜe>lj4O p%[d n!;/iϟ:h]4:IYK2;LY%DS]v:GN]5=x0 VQt6|5-0}c۝ b,4>]@Ѡ gOz`ݽx<.;ߡ?`3!wru{덚?dχ]7~LOSK|0%8%F'8~^LO'q62Bv7HG /ɻ,Z2>mI#|jB?@ BK(Eŭ`I,#r4!l7>MIzVƑnk+;ښIkmvw|?ʁth W'?¬k+|mŚX󵕯 r*zB`Z)bS`Z:`SLF‘DQ\84jCB69#FBc-U*zVIai&0DdMP23gX Tjitcb#pΔ @ػ]> eƅs 罎k"S4_ 󏙼‚~Sm^e}&^5fl=`c .EKxf$ IAD$$a Uh Ok5py_̵\~z1?2n>֑&0A޿1;çY>Ç\QYUqCbPwG1%8,|!^!Kwl7|6?%yp=!C 3z3ŏfFPOaZsϣ?fG)Yٚ$>fϑTS%6IZZD'7.Q{k[p`-hMaZ\VY>,fnoa,5\2ܸtR jD-_瑃v+@R(XەjD˸D@-ŲY,-ϳy†߯Z]y|lۅ-ރژ]þ9Jal)LN$2.MɍuBͧvnd`9m4Wqv_ۼ'߀ar- cYS0ރ GZv_VӮ- ܸD@a> .g!+gV7lU2ZrrރOY9A\Oew-9 jbUZh%'pOTfsþWX % 6z?Q$V3clud'䴹+P-u\_g5HeO?q(Dtpp')zs:C jr8W{ z"~;=rkn-aY !b:p E->Tߖ؀ǧ;Lhi9a@l rH:a?F*<"Vz{" 9)a\ ٔ$.rPO bY嬆V!wG_T 롗闽r~ugk_9? Sd{?EzP-T&^*ўn?x~f:JڌΘ 5)L*^D`XB,NYXOMmt[g߷U|k=vqGn=ZϷ|[bϷe{S^e0^ F6x|7G{1i~}`,*6SNn=bdV)Ɗ-2B 4Fz28S`pɺtjLӔlSdIؔlSMn}PB g׮ght[Bɂ~vV7Wzq:럓[~̀g6)ME6Q8PSDccDFeIٍP@2N@M3sijyM,ODR B^_ q >Z+3cGANެQ !Z>} ڋC`:8fuU$:~\W%DǬ鉭9`ƩӔx}z~ ]mKNH f,K7mYkE!k#"&!k~/:kL[ a4<"j>pb=1h??)ܣ^R V,NUI :o(w!z( ↜\qjh4bM5 4P@M-OA ԝ $:LuLY:Q aH$Zup/?Ʒ=pW ҁ5ZKM = yQr:WSTy|V tHYXJ-6x(Vd%.6O?M,Qy4ZG to.NTjO)D,P:롬svw? ?z8(VnWcvc-J!vEAh7yj mOւY@ xs6ڳ;dDdKE$"D2Y7ʀh7cvEiYإů6BZEZEZ,1k1O*#377k5Z!u%)_ϧ.P])9eBlEA O?k3#X~i"}th{V"7@-jg{rSda(YfJI*ޡl2& 2V[41) *,dm;&km%f's 0Z,_/tBw!+yD@9O$-@조OI|'P&yɛ +* [Tmr{D4bLqqTAXH2g+7%O|Fc#eBQlei]tbm&k]PEśw5Ȁ u`TXJbV([,. ,={s'XWh錨2e+.BlIH^:ʽʫQFH fO9r3}\0w]ɽ+i4z!>gΒsi<}9_}l?*ao_|}ݷ/8S/WoGvߥ/߄7&?^|?8BgoR'Ng/ɇ1?sMȠujFש:Hqqgjȏ\ D;(nHݧ6xr(L!q61Gf,o |"EqYg\^s֣KLԲ++ƫ*xF^>ḊK=mYm;=DޯVˊl%!wӳ]ɗxw}\]w4CD $ϙ{̠nƴxƳ{z/3og؛ wzp;xf>CNRȯ_Q镭u9\@bỸ=+R%(YjU\:dJ*)kYɰz7W3+YWxνOn셞@bL~bd zq9N/khjJYNy[FSik&mSayNNY>r"W@rM.증< Jz'<&EIHg1JwW1 \T?{BSN>B|vrFO;~ ^EO)pA:׏~W_?wE(/ouxJ0WzgktZu^AAu!ZVӽ ETABm1aBdKCb]ђX̳}ʕ 8K{M!*Z,̍FKG Vbӓi|l[EofkCgd'BY>Q'tQde.Fp4FVͥfCW B7H?%72`.0pMѯ"md_PYqf3$![NU?]TuuuNWY+^c+eJ:-Cv_4|pώ~iC W:+1;2*t`HQ܆*k^ЁbIT]$-|_ħ?|KϪ*mdžm0=v<FA&JdJ^՚8RM9QR -꘲IY Z(JlmuC$mϐm"~e}Sd'9Ps*VXM2C54elo-J^{Hh xFR&J`efᆙ D}ϐ CXp.rK/{m]$ 9;LvC؏aˑP$*F, A*0C92fW37tYYv4u,L>fwWgeWoo ?no()Z>IP-}woʻyJjaQ`aqTZ٬\Tj:-tEFG|hAtUQyۡ193<hT{^JfynTP>*iˑ c7H 69z|`+-q\NA%yЇC9iOӱSjO`V'!P-6q.󸼩$JkAD~c͸TbBn(Ԛ X5g?dLޫv`L'>ó1I֚6aT;iQ<٣vR$cfCk Ic ֋!B)9λȜ伉IrLmВl&U;0 J{E~$"V>"Oڒ$Yy0"a+8n@Kޘ!aJ9%&hqkeej5d$*i=1Y:<5w$fO tln줰`$KVvgIJˎr(g2*u;0&^~fՂɻ^Gdp0g"w$GѰ3JO o4Y#7LK0XVƂn%u;4&| p:&;<2G[ǘ YLe'D)ON*KчLd )'hTCTa>t;4  }>Q9u9.mG>/ t!%9hKnRQmNi5lUPUő-BpGwjRF%Tpўx %:'ۢR2tHsk.Z}hG'#qQU Ejh,4S+>GLX1 %R0s&uL 6u@;%+V~S˻xl쐋s4Ż-0Dْ>4V͟ֆh{*@9fOÃn J< cy#]c|U Ϫt({GΈ*ݯ^ݷYvlo?\]}W#W~gw`UiRʎ4 R_7+b/؉|\߽m'<¬ ,4}j:oosJ4ħ9Uz@'tmZoxIZ <ϴk->mn׽웬)?]ov_pݤ|wU?ջwK~\_եIM{ZAv$^pdjw@[)79  xC/'İ|ل )8fF8$vv7LUJ$E|o8P[9'&gb- =v`LDW{WJؐ[U}x2ZH;y/O1{Y*rJ()FU+:p բ&@CVJFY"4"  ~Χ(tP6NBё/Sؗ^:84cJ9mlkihwh0/V-# &ƺL譕46v͝nB]>$HHj~b`{u)]˾;n ~]>$Z4:l|j mT}k1.SAI$5QÜt"x@9!MQ/>&-S=C(˜9p|[nIN;:JI#C LlU1yY8x:]3dFe5L O+?3"xL@R$d5d R$dK 9{cc`c+q@Ä }q3 G-VMl̿hcFWW3EoS0\!(Ve_\Rd^7x֬;gvwB0&UO9n lv*R&c)5ʎY UMG=RRfl]R<}w|Oc6w&zY).M*G flHu*F?Y V:9]Krw^JMJ/ k9HvWJ0.@GuZI s8:@V` <: JTdۡ1Xg:R椐y8gÁ ± g fUncR8Yc$I$9Ow^1|,?")p:$?v/Aڂ[ɜDAi'gqCyg5?< x4 Bǡ1a6N)Apז<Lqx;9ot4ls\5рmNFG gR-jqR:udF,9%8杋NQB1b]ˎ#fL9i1ls̚y<53lՌfg_G[3ؚ xc6o*pY`34p8 H$k 7J}dt[x=WENMk6lpa$ X0kM3T ճִ=ǞxGf-6(Yg+|rfh4I.oT#3ԡNslH9N*O6"t<}C{<{ҵbٖIr.OM[9"H1ɍtY\GVZu65#65U8?=E' {yl̂aDϑ0lЇ+}V}Xf|&m%fx-38~;^^850q=vh`Ѕָ,RIɷY̾`hrq]˱96M* kghG 'nut [iV]op 7Mf5\3I.iDzZ 5Iy6BfH?_wW_M$"LOIsR`j:@Z:Ij~TUPBb=Ga5*ZUS*6|՜:)uW`?'i{T"7{^]"CJkvt$tgí\ܞo{#֤{ys7e y˗go_OGX)]ܾ~wG?wXU ` N*e](Bc`ظ!cOƨM4/D>a -mF ;.@݋bMGG]ܳ*R&Z/#˧jvQ '$xSQzd2I;ڭG T2${eh}J/b >\t2C$vkّ۸["ljf(rOYvSwB"r.]_m~ojH"^~h-}q- Lx珯=vKZj1]B7>K^cIyz̊BTjVlz`05"QpEj[AUDhxK+S'һhJie}Ywp",atU8OH})kR鷱T?8#gfL5pGc\ )dhQcUF?^ CPvm SN-wo>:9cFl/?G`qwvY6߈po?7\;/IPrCE%Blmnc Z"kQŖlug9mO>~ʛo-@V^"}ӊA'IH&dۯ?CN'eyW2 ޛU+Zmzi'Or@Ig5yeHauԡ=$׸2EGκU*FȾDZ.a^{ XS2d[F?MrKJ"RJ`DX4/ zyxS:2uSY\CL NV˙,3'Ǜm̘"wc&+#ǽ.;Q S$Ca0KdHP 5SMnKد/2kR *OtBE^X-zaբU-Z] -.Dg"-4ER`x¡deN5NDnPE"9/굊`}8Vqcjx+<]N]cDRC3^߅wc(p;{.+x_=bR&! HQJs~QJzݡo$Q27ld:sB ҄eVJ@ O!/1#;}\pU5qxv{ۊT yd=!ɄȐ#(T 0&zDLkO?X F%Jq8cP/rUGf2~lc}vasDACtgDaNsrϛs"AHx^m47(Zus"R|^궼k# q#YV;j'{J,㮄-ޓ?tz l='=>)8xɜH,Fk 9y>~~F`3jl<9\::)(2ct 48F[PhO3lmEV:ѐﰮ^x=jy⚅F< pxPHOk0T>0{Ge}$N IX\_RH|$*gǛX߇HHޛkbekbr|_O*M-/kpLP;fvH ;4ب b'z*r`T͊kUH5b]jAĞD,fo݊JXѧш\/ͣRaDlfj=ou/R4S\(ګ 2>^>@$NQ# EJ%QJ򔔚w9D V"v|ECrXSwN8fBG7+0|z_mCy$g{=Y|~|p &Th9nhd( SDFKLҧ70q|4,MG5̄'>ӥ3BKٴZbU+ ۓlg[_oш}jJb^)0Pc<|h1\؂N,BL{a vEHwYvrç)+e4a(-%p!o? Ӷ'2v3ABu.8Xh,qQ`&sAZKg$a VeIQ(>虐0vr z L$,,+s$)R.$je%T҄HcKhf=#huylʨJ`]yQBev c~.kc zJ$jk9 z-`q>gwZ\/M0ݤF(KPjuGJvsnUVWu9Z v->}wa 6T:;L+q1%X'$i4-(N3YtMEYL$L" 92 + ^u;B Y_љ/ӨR INg@"(c /1v񵱇OEN ߜo~2y3@8{9t/rλyxC+|uEmẘ`WC~rjp/~zl?/*+YY~xŃ= P}2Qp/C"T^jVKǐ5N79[&ٟ]=͠k̘Fx 6)oHM=W 2zс<|u"* T\1!P:!wWitO ǣ65ttKiy!>-&jA_Ҫ UTXGA>Nte"7"R@b34Eb)`:䆊.* 5W:Vo.9cZ SZQ%UdG"$Ҙ<%Zi}P1.i1!c5VE&p.|IӞp -‰Cyr@‹Sw"-;:fs(Lu% 9Z(IsOg_DJ~.h6^C=/1OILYU  p`i./`lh1H 梣s3 Cq7LDpaq%AI5,&ȶK1EE&b25UV0CJ0^>FYJ$c5KȓP81duˊ֜.8V@B/:f+ΥLS`<7@WRUQ.)!T@; TZYZVUj{ޤ jW>b ]{{H`83K/Af>m8AIY*3Fr'= ^ZG9)>J{>3' 9rN:M TA#N@72 `|jҲ%S}QUm1ؚO4H|2h 1qf)Q{cS9{~v8 +EiӶhXkNȈ'Iʉ g˔rSY]e4u'(HZ{,^c 蚫;eڻQ=F7pN+:G15Cs^wWk'DgKJri(2H1 5E!Xo.o.ݝ<@BB M{D}*WZgQd8'Bn===zS\ 0PT"$OIYIݧGӂrcpKI SAd.t:#i<˂kݧ3O[pk#gp.)uB:.2;8.9+bteyț~<]VTz 'j@Nj}N_ӌ!h5_^ ذ`'fRsu/r倁H:(KJ4JSΐH\IeM\8Z.j- HVj6>߭76p }\7JϠӉFsCyU+!_d-^! ԗI*$JPb](6?p-zσ$nӚAXq灼 Gq9Ɗ2UeH ~[[$`3%7/^1FJ_Ve`~ɼ) {R?ٶ^r 0TD}'o** ΊBs4/,H"9FʌARp_3jj^`)fȐuةC[F-3t&(:lFr˯V:!/x%?K!r.2YOkn=1pPPIAZJ BˬHJDJ *3yZ&PK:p),$#Xy =bt!Ds6NE: G1NXuJX8y3MЩzɾ?yz9=zGW^)&Y^L@5KGA_gCH@̓GoVjRԛ@D=؍1WZ'$*G߱BMp:R~R%+ #.ef,f`#'S;ZS9Jr4QJPrtYq'*sA$7ɭA|K9!B/e2MaWُd}>+ 8o?Olʜ-J(ͯidW<ᙼ)"#`hL'eٻ6r%b<q&/1$<~IziAߠo{}aYjrƷgz?͋VO_bvy| d:|Y(ti]1"s-PP7coK -:zz^oŰyMC{K_ZBo>񁅴JiŴ,ORt _3bK)J;D$mПkevuhWW{]e`.'ѻ-S?x/Q.%zXO Iyt*07qveAF/OOyD^h2ifǸ2^&p4i =?|uA={7ާsrϐa? ۙ6j%RE 4{2]1&|Hst. )l!|`sɭct\S| v5شT?_?rasiJm)纊AJH1 `BD694 Έ0{CIc@Zᶶ~4-~4-VM! W` K6Obq8`8X`-Q0 Cؓ5@3va$,0S:s :M a!8f e=R[ AX=iAzXR0FhKFecGhȮ.Z疉u 4v\и2P Xr*GVJᄣIRi5=-pÂYo臮i>M6g *QWT]r#0mwӚ.'eTtx0XZh lE\InH.i. h(/Vݦ5RݦYumqN%Z_*UdѫȢWU:ѧKKrnG| /󍥅ŮŮױU T!&uwB<:fK$I},&Y7YS63 ~ZvYд"""̪|jMZ#4E%ŕR4Ug7u]R,R8,RJku䱁2. t< 篞L+w&Dq|VM)QfooRT۫.@86[:6US^n܀#y&5{:E–+&ܺN-wtB$P9PƄ!0MƵgO8p书X52BPxBqf<> LE#~润(Y&b&bt!UFɠkNYN]!Bn1@<qq䔑1܁;W*p4r`1=Մ36Czh<[.\:fX걳9&&ըqFP&Oڭa{ n0{]t=&´ CC{֝MA1qi'Jd2DIdo^/g~:қƝ%:yӟ޷;-W~kfǭ)yu} F0HI8(dȝ6%^bMqWjqv bن9"EF81U%$ 02J@hW漨k?ge\ZL!D. ɂ}I"|9RԬ/' B4eB'Wr08y`s=,S7бjc(uiN3B} ySHYT1 _ٳ |m|MLR%1B 9ƅ\1D,*8s/aPl/YG &kM J&bDյܛ98]sǑWPr%=ݬ}8Iv.9,;bkvfV 2(Gq@.^.vD(Ul===+E) >]^L.船1h#JTzc1M$'x'GCA$_SY%ihі]4 Rh*iUګŀ\@_Ր4hTrThe)-f-ݰaD=gQ r!q(P*0*\ [12]k G% E6Jf@ d-an)N""קAO.æԮGnA5u]7|1LHQasTգ'Ӂ8֣'MŰSb$+˩4EOMk"=OK0#1;W],J!9FǂX%8F蘶dsuDi|F RYƲ9E7m32Klv,-Qmr6%tL*['0!%9hR)v\iD4mag! زh'#klECJxWW_#;j :C]D=?k{+X,>babdb>꾺<U:zt@g9m;]V>hE?&k3bǧɴ禟ɳpuHb+aZ|Xy}t=ZXp2㾕_j桿xM旭#>..5c_::uHO7IS?9ǯm:?}_wu]mQ`dNxx74sg5)HylIBքPBvK-*ɄxdNvzY#(B*dj\߼~9$YV A5in9bC%*6'h:=rYϹV>O`*Y-8zcL \u,-/p/,꼨dV=~cxku?Ndv+_)J=JN <[$HK>X2%) ;N%z[f3ѭ]ܹs\~smJ[mms}2Bu3kem`:9Z8=:kAOZRF8Wrn0!9ij ;d/ H7Z|3'R1_mO=}n WA`cnMϻ[&`DzW{l m L|ɧDusҹ@*UW>;`x[,3\4}R ˅C䔻U\*k2^M<#\O`;-f=i̕xReIFrjJgШK .JYR\4M:{F$j5nTF7O̮riѺW%yq&e@U@cwy *aBXdG-:;trέF:k ^`"`'YSEHi6~ V5gA&-EuHiϞ.;7gmsTwƜjhJ62?Gـş{վ{d S-{ދs^b%-Q/)Ӧ|HAog?sjp4|i{\_&hNv8.ޥM߾⻷x4>fd/޺7ߛ@gJ!AK`.S>01sQy;a13N F #u/9pEn*BL<=(.s|T[F:!I)zM#yFR1Niۘ烈G?_4[M_?:يEʌϬe ! 7xBnHfeG!  0AxEϹtNpRm WyA)0 NܫF0C]'d{ml>x5S9ËcYM!B}ۧш\⢏YRq+SIAdr31K~צ2Ͻ#i'EDOױ"3$6񹖔ƲHfĻ\H,IxMEPښ~&ғn[&(Z*p{bӏo.}tn쾿/Ҭd0@Fѧܻy'DRLjk7~3 >=!z]@iv-jr̀ ¬.2LQj/z h7iÚZtW XeCԬ۩PQpf*V m{aIuU LǨ :֚SGG5`ޠO$q6αݤ~cddԎ۵:DlwwLu{px̺5~Ѫ֭ y"L!YTL!f-k/lȥ?3T$\iih;UoqUR@w*dc{0#0ˎj%zڏVQõK/ۥgVVpvmKC<8LF℞@^&PI;$g~}X@ۆ'.UTX|yB^>0 R4nm{UgdDjIFR?,v +[h/=̈́ntg_G$YQ,t`$[eъP{W,>aj˼J1J%r_͠Hj EE꒣dk>Q{e㇂loFS5p֤LRh+?9FU2)$|On}orrCa793 fiޥK9%.4&?\}|%MO/"Jw\H]$WH+S,@l>8NI1xl0<*2]n Ϥ> /3;Z˷4}eo^REX.駸C/a[=hjPs38Z|3f2K妉VLȟM,zȤ-&d^ WiBUyUY6iJ?{ƍA. E=w8-bl6if۳I{C+4ZV``O+XUGStWb}ߧ1RNsqg-bza)8iM_H5}VpcV :JVޮ3A9=)JwAU /]'M[$Uz#f@ &3rR̫\@\))2 *HF,g9ٞz7p}EAv(N1VT]e !y:U$,,Ֆ[\ nY>=+V:Z1]bjj%5Ar}$\d. QFee:?{yL*pBTvAz4deZf :s^U ½'g.eeG]*RCO'H7@/B?+YRjddjE=T5qkiq(y&$ LC@BeJMk"6747F|#to^PV~\*zBq=(HCKAB8qȚYɣ[|\%}Le2K7&>r@Q+'hu֔ rca#zFȮy:9uh91NxP,C@юMD*P5+Ŷ~)&8{ |Y## + 9>hC[䛒?NWfL4io&$X}E8tIul[SM X*OE~'"SIu0EH"Ȕ$hya\nU9`k6ȡm2c3#g^X(9bb*kR]f%k&H5u֣X9BMq& %V+~n rQОz=*񁑮p{qW#ܯ~9!5׎抠 EB!s^wJ(2sw\*;n~=ocaӨ6v <C 7TZJ hXRZ!<:~VD^@W03):/cDz- 6$d8uud:.*!%p4U΅FlF3e9Ucrx=ȵxc-X)!v2_)^=y; 9Ex$;PudEˑ5@u#j{{O/ Ʃ±QvYRǞ:U oS3o\ko>T]L\j]L-&qCcᲓYδϼP j׵#Ԯ 2wU-zSS7tyJm)'r.'3<O4gY+d)f@jbefCH~Tؖ\X$G4CL'Mr(UI7nȲ]2 e. 3|sV8 yWO|0zAH^iߎש\ I Sz5"DǐwSܖ0-CJ4M ]62 Qy N ʀ+xC*xK-AкDmr&<T4b46Vm hMT!.}޿NYPHbbۭ?yY~FcҘ]\`cYxeiL}&%U 1J,*{gk$[F?.SZ;;q\Ӄ3;`+L@O5^XG_5"{Jľ+YWtlE)i Lu3=>g!cIʫhu4FµڄdY%_#2fiQo./Vߝ2R$āgHM^7 ?SA3*o;!Q ⊰x(/oFED|zܔ~p7ȇjI^qT*)FK1q8$a͂oFմEk ynl$KF]5L hUe"Pd#}q+m:A *wo&ׄۧ/cvr /\5Dyn9-%a @d 8UYPJ/C'zC/@4!gJ[ERJI!M2Ifsw2^;> ӥdZV&7ZR&16-q*p" M9έ33K(e"4pRf"w.$A#jOb{-fW|d|co]^l&W9C33\x[nUJy!e J  GA<۔"jqCP@tݧKwdnx"zާ>C# 5.GR~4ppQh8Z a#p|րP @q-o/ꎆ- Wq?kD ʏ'\-k+DjXq.sNg|׻'q(vH]] U7(!%I M[>,S "Yg cy.^ߍWLX%YcgUf%hDX%&r4`ѸhMF5z*;$ )cXOw*.9b޲9֮jQPh%Aϛ]R`$割ޢoubE$TJQl/s XbNKP tlfQ f))6G>kG,xYwzތ0b}Ȋ=.45qgϋ(f:/CYhQ'G]-xg MSې| P6ώw*uNkhcv&3(r̈c1M:ʭ:EWiWcMN(S՝U/׫0΄p7]3g'ø3a]ӡKq" t&ѯԚPbhE.[љ=#KT 8(\TG23>UkZ>% m6OBWhck)z-=uLs\>jrTD4tnWn%i ًz}_,79 KRq&ڮxvet߼ #b0f/0E[_כc8ib7FFpBHZa'~LٚPlZCCU0@Kd{u?^3fcTE4t2K|F%Bqʤu2~ "PM3HVY;œuT˙Y# rڤ\*!T1L@TS¤z@XK{ӌ+B|*hxit˂^Caljq>MC?ڙ~ATO^.k.^|X{?Ϗ/#eUW\}qY1;|owT@xMg|4)~F)~߹Og\Ď>O&| pm(WiJkt>= Tr^2pyC%%4[c~^aBR`.__o+3rn"š` m jChw'`?׃P1@h9|u|:d-<ŻznX}rUz%1nF=lҰчOw;?Z_wg|9q .޽|;^^|> Cӱbj'*q21hMu_δbˮֆѰ۟*rbHq6}|[J9$>UO;9YHQC9X҄Ee'E:Wx{ѨxdâA?braLuiIt]#IspNQU;ҹS4/U t\i_MF&pH.uZvYz2b@F0|kY"ҳufT8`tx~=Lp(W-,u6h-ϓ;T5q:KLN@Uy3T) jA暁HzR$kqpf%.) }#:]ڬձmEk.4I1 {U XV{|W+JO0C %\O;fOͪG6>}Oreff[vwKZgQE9zx[y]j!-o\T*}*gN崐f s*HM1gO F {9 ΖȂ'5tM+f@Zu&p>]Xٴ@A?0-0- 7@pS"c E R2'3ͭIu t 6@Q?|_q1_z%It1NF1ӻqJPC=c?m_9\w~*(ey`Q1sD,TC;1DR8Q쑫 ijF/:i>Z{CK-+K(h p*aKX?p:ifmCjwH}=jG-֋? x!}XNV魷*𠢗)W!9O &p 2ʹӝ]Tw X B٤i-\J䮸76Aj!cp[cltmBQE0EC^S:uљ]l&XE)V a ؤ[ hs% Da^9JT"O)yJj7w75_nvv7!@͈ r[M3ƑdKPK%EvKq V0 W]Gd $,]PŰ82j? %bFy0X ["AP6Ɋfk5z~~T̕E9@.u'I329E4+WppݻF[)rTt6P_ 8ݷw+hwBCr-)I#5&)xRNm-nn]hW{딒YM-x.S9oV ?y UN6VƗI8cC"J+VGBNJ4p˼Rb+Pۆ.R1#WVcCb$f .vϧ{Foܮnz$(p3 JoJu *Oi}}(8-7N!$ l^>=?ݿpc&ч뫺h FDZIJzH(vS \>,*-KD"͏`']LA ߌ?7B9 Gԓ>HwW-d)GwWv[Vʤ6j3UBbz%9g['}t{ԏȨsS7$l$9!lNqUKP% 9 <ѩehyƦ]?dxM }_ӝ%-cOK*.?~-QeЖ1PC"~55?y_ R (%p<%dl!I-9;x mE d$>%h;HA2!)DſɈQs_VBhd h|&,Bt- pǥj"g)*Tv&>L}ݛXv+b ##ى:g9+SN4%jMHXJ*R: ,a]brK,N(С|Ή?TL8lҭCҼܧ(|!J08'2.C8?ַųa#>'rEO[azlqJ™ :~݂ek^srAе|#J5gBfAA"AuaJP #O';1k-YdhFO0hw(hCHAa@Et5Z2xGn$j)?׏mx@ѼV3TsNhtd舥W9bJ[*&zՖ1+jH,gܧ%6m%B,E% ?Nqz2[W_Wq]„"Fz@{ !$VDo$mR7:9'EF'5u+wۯrԁ+-1(H_o>ca[_âsh= "Bs@LY`z$~I=:G /JɭRҔL"`Zj󥸲y=L>ͭ_"ILG^{&VLqVJ :2weU ]x/LBڢxײ̯,9)0C֜r84LEXҾωrW\{r"J6N$ e'_,jU>Tj͋%u )( sX+T/_׌R+&XLԓ5QU.Tlv 2Bzƍz<~cPdkުvy~4FbEg=/"Y$c~WߎíI#@L.AXְŇnON۔#X\N]bֻ82~˦K`viMH4w$"z=:Ϟ3,н#DE;v0[ʱ'7y  ѫCLY2 \ƞ喥 O%+߮7/_tw0 mpsH\慓9z@߯WI@݇-l8`,'va[ s3OI+JTI)So)N94 X S0 SٰYFA`Kj=k4wE r!iMHTN" Q * U8:C|MK~78@YTŰt5FT:p׋c,f9.qAP~Ho>,.%pu;\8ODko9dw:K^ Rrk:Eه׏\ ) z% ĉgXt1Kdd9GD(B랕jg5*k Z›I'7g[9/=ǔ!"x#Q2n[Fd@yqdtMX3Ŷrp) !-uW;Hh`̖؉1W ([qy[w7d/\uͺ}:m$4s $J=vGAXاvX@%RA .`mttvv]wYG ϔkUisF EI^q44RI,qN=}t0R2U,' VsA!y R)H'-3lc%zjb9\oY;;@ 9#{VNR 4ӐK*WZ4z}>7L^ë34?.?\9X͜O]_M&iB$ThSr E8y}#I(R Bmެ SKKGqҝd1~g*ñy˺fnٞ~o~Hd<~uvޟ*7W7U).=8g'vaFw^PYY,0~K'^Da&M) d#V8GJ%)yr-&֭~{ *,!7jq$sdӑḁ{SPpybrJp0^F:aqX#I9GTvoJ{+dyx Z›9#Kh >:K>745t'Wf򱂿+hjZf@7xu4L[cEU)/IW:9fҫ3a6!%mX$ 3b+ƽ4B2)~ N+Ix" Ĩv %FHi)cb<mQI/RFs{WO^5;i%+9'j/0A [ hOh1 %(ER1i̮ D_/KA/1QNv>$c+d+OH![5) $}2ƒ^[H.L# dn186`X,≝V, wDgM>V.\^Npk\1JGR:Q%A^%pHyVS y/ !^IA;RɶRqrԲe1nH~ \`^B AF$Z1i&*'%2)yFBL.N0ke59 fΛ?\s>>)uFrzR" )m$hJ =;J Jn _TIcXYWYV,~Rb{Zb '&R YY b6P5%VG DZZ'R<]jIݛᘤr"0)z=ؤE'֮m]|QF-gtZ1Ys}EW xefIz=9mhQLNnr:[O:W\]d*6jzoWn5{z+oqo+X32jBRj$חll>nWZI)> qZvhӳCmٻ6n$%wW7rl*)T3Dr $"0/\S)1_~F񲒻V3Ƽnwk]OL(t]]CIw6u/ϣiL ?s:~d׫W_;LPջL^ ـ%W\(ҨMHq= մ|Q`nu#2Q1@2sg`I@_]4n,Jb\ "8CAr.0wcАB4NfUaO=SQfTqM$2͊<ڜec:J4]XgisSH*, nS%ӺGkUHE$EZ% +C%OK8@;NX#DG*)T0@ u){i=Q .D ;(H EY`[l$Z $8!MV T2L.~vDreh/Wx)%uy5+!_?L y;OzdJp>BN1ƴ@~%s={ ?3_ط>Oj2 FpT=>K&ZYT ADa>ϯQJRV;۬zӷ8D]Ϯg+AXzM$CIu0΋9N)BY;\`yA,rb*R DJHk VB##S2{Qm[}j1H$ <fmRByƐҢD\YȔpXwBm:Fg7zf:oG=^jjԛYd^zyS2{ _}{#^- 6H-]BaW^">%T_Oj2w QrY_H}W_=x@s.?QM+#0:kzU:s9p9;z$Lq JXAՙZԗO˧{gf5+(PMt&$֮2tD?'|OG w$W >&֠ϟ_n ME~[t3!L_6#~Lgqx7B^zf.>eVK{XKϿ|Pݓ[P<v=y,q݇m[?Q҂fܮV3~x!LS`iR!ݳ`Ti*E܌-C#Va!s: v &"{>v.B=vuB쒴MvE,M`B Lb:srV-v?QۛYOY[`T:*D{M'"m(vu8`k' -Fl8fÛlZѹ !kPw^%?5HcD>_~~;Mݰ{IŽegu&9Z#Xge=$ Wb9AhϚ S7v6) b{c'IrPNA?SKtIʊ$KMX$-2 rHJD'U !Pw{8~H'?6N}swǁ5qrգ?,nVAl97D\|0[}$سf|ӽ3cJݼ2̿Zʐ5UQ)rR˱.iO'Z;[sf/vu ^ݒ݁1'[rǟj]-`Xa_aDWt F4jVK2($Y3RZqp;Fy!3׌)EKe%:XKB"Rk#9=(:*~d{ o38Q+.аO J!:U)&'Ιę(̐)*) eb)ZhX_UYgWr -yư, &5ոf5 Y ktV8ӈ#yN4Y&)*Hb>q!ւl̬-x/M,N@cؽ f}~e#wtۿ1c ~DNH5&~Rݚj_ $A 9 A'q+ڪ1$* mי#T]ʅ&O ;6tNcکq, FO`7̟]OΓk_/9-UTM_/._,o1cZ]aqW7#3IFyDYv l_`պŗ?Znd | ثTwy#rzn^eSɝ!_hZ krFH83 aD<c؉;T[1؏vXqs_DϹ}ޥm5B}*qG0@4YLX=2'*\H"EA-BƂ&2m:FbdӜeS fa65ŧQ݌>eJ'e"ж|(=-w{hb|o} "HN*1NGt\_q Jztv9I})eHw¾{PɏMhVH$[C"Tuk$SrŠI2Dkk P8u ?9o} CϟJfؤQps`M.t{ͮS$ mxE"Pc`Qө Յ B\%J ,]dEaJi+g,e:EJ|3,զR(!S0sf20J-5jJhS@fzf}y>/eMl$Ae]bdH6BPwwvU C:[jn$B0e1F.2݋}ұ&Db& *)ъ2BScθ+uf!-cJrU[lBeC3_"VZtLRőN\tc+!tSreNQNbk&HX1&7сG> ỎqGY7rZU#ڑ#.d~y(ƥA_h)pB 5bFB%޸kò́&OKҎ2b~ǘb/`3e>]N_ p2c>II `f񓌮jQ׈d_ID2 yfS摻SG`ì]|^]rq ]^. JүQYrʧZ%'Geb6qo;{,-'HKJhiz8b̨;0C[ca#ulT|rovCh0F5'Nh {{mI*54T XA.d&v?K\k5: " d0)R$ULEeʼ`VZxF{l,]DkM!6p:2A<1 9Ajgj=M ZbàpU- ^قs6 ^ŰP~fz{ W˧{gf5-*[EK\Hp:c}G@$$8:$-d %tź2(³qm׸qs"Wу˧{b5m99J^?8Z8ǽzg_ܪBlÊk1Tn/ڎEȐ<Φ3jQħC(*s4B*1o5*3sEhs%MA:Cqej(&=v(P}-+rvUF\b1`:rV-{ʙv's}ewNwJNj~ȶ_tM1 s^V=Zb»{EDž|EG!T< 9ȊTV" %VaeI; k'xJ±LXX0}El0F GR{Rc&n!.b^-1K~쥋}T E ;[:q%Ӎ\W\ SˇV?tᢥEdV@]^\TFwE e+dkºhVC⇢Kh?mJ[Ji 1iӹC&9\(n34HTjՅN LS>[>Wyd)T4\!zWʻVLBai|hbƩrO1l%3u]MLÅ ~y>rW6&`t{p^D 8Dw=畲/4R |w^!"Gf|R7 i9Km,CQ>v~}HIO[E-ʳWRUT w'c-yW8c;V|`PRĭ۝S0 %A-(_iV8>u#˽u9QBn7砠c_%Uk-/\/yxS54Rj3#d[GN4 =FX{i?wӞ)M}klE♃l1Ax&`:۶'%}ͪR3^̖AФ/:C̱9n[ O/#!Pqϸ{z2z@rMIvT=${=tnݞQ; DGWLb=Ui5Å8Cf苽ya4iL5៍w1Xm/,r,rܾ cCBNIEQӄI&i=IKA #ϛ)^|&EҿډjJt6o'9{[`U|gF+˩غȖS*JJRd)YD֑ȃɘHpJznQbO$4Z—*"[`J$F9IECYkB(@H*ͽAtASecZN/6_WSy8\jUAIM4'M- ")AzspjR=repǑ :Z"%)sBZBNnSuY ҤZEWlܝfXczoWrgdP̢YP8gMaG_iQ;N׋Q^)e8~\|~z`2|GӋ T6@vFۻ)i?#ӟw_鱥8 > |DیД~ËP.(\xɄQ.aCyK4͘*{PFe2;h)$[ŝKP"1ťT0*q"4&^.h~icm15EVzwAS 6q]EVVOqk|t7bkםVUH$`ٻԈ؝f͸+_Mp~N&.Zem:֊Xj72RV)ݺzg_Cd `g){,TNWd\}MA׶q ]cl D'͆9s1lȕL6oa9MMBi"l=pSyq< %ՇVSm] __'6o~{_a` lQw;nH܍Q@ՉFa8oOq@v!On&̱hąXÉCȨ4AK` s0gkqd֧5ZB1OA(anwAZr|Y~jG"U{)O \os01ynƑ=ѨAyTHq 8wQ+/ 6}&! 4ĮdBDqGuVnxи{&Tc kUc"9;iMeVVě s(FoUgN:dv?_^Gsf?>|z7=Q_\wFMHO;$ޣYcÿk6JFo"{g H#zL Wøз 6{޽"87ޗŶEvˍ qVw0F"p{քˀYHd" ̢#8_ĩL+f* s<ڳ񾵉&6IIqwe!yGlJD7ZXYGNjJ=}-6WJ&X˕7g)6u[g6qQ;F۹I7+ h/m_E8ҿb_oHyL.N_Pn rWoAimCҶ 90NԠ:cNm#oꮂhEߢ~U.pǫ"loϧ>̾NҚݩ/tmgb b˳n0{ dQ$ơyfvP 1"VrPR"5CT*Ju/^ fX˔ޭbK_wq|ЉKT>W΃i {k ~2G,l^񤺞*w<60:}|:FWk)Bo>|,s=q7z~mtt*bd $MjESx)sFi GDUQ;y|v/"qi9qiĥ|@B+! a,Q`-x AU2]?j'+p~LtHqtHq;ҩ-Kc%$)CT2@HplsǨ"\2HX>j'wMyv8Cq>CqP|Iq?"=p *WADZ ƝR$\ 1Ǽ87 $܀dn@ҖEs 0pI(ZJXA[ozSGDUQ;F }k_B šеL>t>}姬%= s,NYKh=Qt]_:V@h ˟~C@^śo>E?c#>q.l3C?_g2Y%a ^MfoFz^,Y؇OxQI.*I;YNL~(T<( JdPn8]ƭ$rZ 7:FV (^hM|pDzLR: $a5X!E7i4hG+k@0V8d(Z9e'ffjۯJ귫y9-ȉ)רr^DS0׈::t[΀Ɉ-g7?NpiMv~|ɢϋ^TɖV:Á7+bZў\yT_SGpS2p%%ޑV>]ׯ!Kp=[ō2\3b:LA_܁H9C,-.KhIdRkʠPy\$2R8\ (*F-PP)oŊCw%*rZ$=³$qmė77O@G&zqjh-삲>VUSh>c 9`iJkI>ʴAdr_]s:˾ d)4\򄻊IAa[I?%MT0aZGfuP(Ks "iPƀ`)"$jx!(q6CGpD*D 7{S 85c$#IDP\qzjlJY Ш -mm$|;ǩs=0G!㍧nL:}@(("Hd~8@ksq&ݍ5ȧ7t8⵻7O쒢ƁNx׋o`\cPH}}!^1񣟯>Dçڴ{g//Fn΄&o_\৤O߹w\ݾ+9NLLڦzy/ч?I.J.fq|@ aۻ$kB{ZDFY<=81qW,50>3yLhnp=l7 6 oGh@{5 ֨/(ꝅGcJ (F 1t jAD@c-vܚ(u!W ITUfI],/ۑ^&R0-]/eؘa'wd @ 'Lz`94h胋"P)FTB1XͤuLJ>5((94EGcG8†(&Bb cZx)5Z=.#`?Lwmmz9XlF ػ'qriaf[Hʉ?ռH#^pH:ؒ=_UWUWׅ C& <&=&\kQų*HĘfZquLY@!9c \jֆei~3),1+ X$QKD+{|%(0^)kWvP'=w_Nqމ~e5 z$\ӽs4SK%juzі [r4YT>xݧsOO8zw{D(M?fnYW&s7Dc* z|*x*{Ʌ@+_m^HG 9#gxEQQK#NN,Wa27>PCY ϒ20Wdd49W=|\-]%fO.O$I-޷ic$>c|f=JQz3B(7c\CF؛U!ҹ?W^aspz=ws_XYo+bmt\?7eйx؛C^Ol ԮD\Ҟ&ݚvgymVMGrյZ'YĪ Qe֫}qKN rMwur 8 ogx^]eT3=_QVG^.x|;DQPΊd܏/E5'(%?L&w^QP ~!ګ ؋Aګ >q8D2tu*w>_KӼ 5uBi 3CV*;2J{J(9DV 3yKy} -]NXpx"!|gWԘsl;!P%qԓqux]\s^lQUݦ+I2 !6|um*SȽr=EQ(nL[6A'I$(b"&) gtӝ׋yq'ë4V4|-;3)BmXMe_3|KJg:ƯTq4&$N`蚇vLdyhSW<ďOGװiB荆4x~ΎDӊ>![:S7m,;\9R֐T$j֞9Va7BtAk;%W{'-u!W-\r5lwGXFUZ%HЎєEYC21jszQ:֐6a;jx7rtAkU6ڟ܍PDzg_ ?d \d&p2[sW v*.`r|M5UmEdFj%cO!{f#$?xW{}`YK!Ѷn]h9竏<ЬP% t[L+km/R=ED/zȫ%Ow_T1Ԙc:ȥv4Aiҍzw ZZ B*8%cpޛ#O6 )#,8͓&1BIDIu&*`x̑g/rv4ZɡGlŭ*YJe2Fr/=*DAQED0%QL 8uTx)Qmpӌ;+k ET 2Zh%ssPG\IEXpV 0)N 0.c}dPZ`4x*g"&*)7v0S9Gi]O+ւn2B$u6R S!>zgxcS ~ⳗ&~޼^tƒ 5K`&W?saD Jkӏߜ p4I4zg93Z>~zp}W# RHW xyሌI92 (-5=ΡGʫZ7Z *Y eh§&swC΀]7tmgB&~f^Rnak)T Ϳl}ASU~i^js^|[{dG_T{-w{Uλ7N'LX4$ŵTDJ֩|+QZyV bC ռq1x3NCӔ/=|dܣ]q6g2}`!.) Cƨzd{[~9_;q,wkо\myr tCY!U.VDKl c@qO hݳ.@e!:.@EonP3*jt}+t=;1ON (Ӿ\i f=oJ<ÎEAEE҂mC0f"ۣr^1Q]`҇W4`6yCp [? 97d',S$HKji{Kv$ t5ţ `Pw8/FYB,?'.ZgK@R崎*:5W$C۳gbA`-V\)"+{-(S2p F5&< LE;l V(rQ~ "G"$ Ieuk$抭'bMIH2m.akCU(=hDŽFUW"5RY<$ANSE(Z%a=1Lm6cR=N+.jw {bެ-% [_lCtL Jm ;Tې(OuTj@N&6u\.(J$@<?7(ۘuc*؝rtoLD c=~9CȄz*_.Bw27($t z4Lbd/i i8ӆHw(Ésܼ!!)=`fLΤ$9max䨩$ 7k;¿+ƊJx_!WbxYM]1va,h^i9'ݲˡWYAAYsC$4MEP2å}x^5 22y޺Go XCՆw,nCv]= PaGm̷{m(<ÐfHS";iǟ|4y!zL@1-8)PtC)i˃uͷoN,OS'g_o<O rvvAdcǫQbCVj򢿚lݙ/YY#e/9r, [f:푛 _q[Bxƛ ~;CK\"y絪FgzF_hT^`;~ V"KD6EI)b/"mu}u9:H\&2G_.:x/ޑ̴nTh}fϯ0,oorOE)ޛ7wۉ}1r\FR6@a܄ήLn(&E>D$AD.B^E2JF!vd2tnDRPfdN@Tv?Wỹkx 4F {0oneDePu.j)[?T` D|T4n;bjU6P<(L9+u%`HS< JS@#QA8f.>f&F[ R̓ŨxtƮÌoSP.)Xzp'AuZ'Xi⃏{)^l㟪_ A J):Fnc)/yt OwϔoݸmD^1 -vvBB_ݵUu5,ToX)r %VhNw Uޙy\]rn|/VsFьXoFШtTeܙK5% j)b,YG^6UDwU[OVM6Def߮g/ ^qO_,h1 Z|-vOWӅG5Ho *xH,p,1hMB)12ɺ' +Kg>lGw?ٖIh@z6k%}֓9jBKs>ƹAiT{! łRei_W$|eR>Nq!"BHrEY' J88Dp%D$"?1UG+'r^ݙW =!%XiO+kaG{' O<)icrP"y"0z`Ĕ NAHǐdHX%X& 4N'':]sUCXԆμQ+tp[1B$Xg (~ u>MD }@@"[7J:$< @tӖ* V3()AA b-)%FƔaJ@DʼnH%a/F(Ps!q#ĕhNV@Q"BlF\[,ߔ+A6YW͡#M5*)_E}JȭV1$JZlė5SB.萳m#F!"UAQ%;[5nR>n``ܧDH; >M"V,)֦ޠ-x& Nx |qIl}Tj~4 ~хeM 7>~Qixs)t_E1 9Fq ].zKpgU5OP@ӻ@/h1if+@9sLf~"裍Y )Hإ0ƺEfv&iH{i<w6 Ϣ@0Yg(>8]&K?}|ws_ͦnYb*ۍ]S)U{*+,48E1 (4oTDb"oܒPƿ?8( DAI) 3=b `@@ 9 UawurHH· ,wD% (1D([АAv.fAQĄ1B OY܍9 Dmbg8;Yg)6'oУyw ݖ8U- 8淔o(=0b?==˂U w7ٮZ+w 5,py&1g㏦߁#Ήt=WAL}LQd"!,ZPKOᬊ|ʧVfwLmƼ<ߴ&ernHKu+S6.(堙V:C^Z^@(.HP+oG:\Mu2=Qry$'XNY(bv\t^~ztj0x~?Vl&7kJ}fkoYS',>&<Z|e8^/okL?cQ)hA~ɔn_%bP1 Ữ;?ICmr9y \PLJB>0,9{YZA @bXg2"\Z ɝ#og"3B+D5 츽;h =G5W1h ϼL(ҢGĂ˴\Eh b׊t  x8 %y"G}s"&+TΑd2kv%b}HSٲl.Gv0ºh͘Ebaw%* MRqaY^^Ԣ J<.Vӟ'P(QΟ>RZUqmܗM-w; 1c4 =kuEqd Nc2LN ['S U!#j_^tj#մ$g/>se'T\pȠh ;Yk8D^~g_)kM%C_O*砖Z^VYPRf$2Q"Lfe c?mZ6\*DB_.3a Ӑχm5T"A8K,P qp6aTZ"~(U"CD\a$^4˩vw})(D]ؾ/6c.¯#ӂ|/b]W+Ջ<!NloeE`$\G Bl0EvJ!0Tz"5$$ Ó&Ʋs+sBJxVca?ϸ?cDo?S$҈1G CLHtS19dIP-±OQBWΫ$%'̱'bcDPFjz].3!% UD(ż_Ap"Iת F9Sn!hz# upf7VrbVsyU| Y2(ꠇ"HIRցU"#MpkMiMj w){DK)]Z۶lZX9Ö1I`AȃHy<>k~5]Cj@R*Y6jƂRE)|Aa_FLiLs_iLWTe;;N=s =S~X.u!aE~%S}]j/Kb5d;!-?~xXys#o߸g#/'KrSqToftLr]ڛX-($HOx =ũ^!+_LƛZ^e8 i?LꇷCA/3cDI^QCA@~ȣycHpϗLx4a!:QeQc2D!5 Sy--T`ǀ=Ɠē2V\ G+| /Ŏ<-Cbݟ}`>J_o|H`U҂G=QR Ue]zu )Z`/KճCD}ebtr˳[JW nj} i3q"Lv1MiUs Mg/#5d+UD0ގj{5Uu4( FW0b(!F.1T@:W Jojes.k3Cu5>8r6`.hFA# @⢆D8qj)>SeLѮF&}893,5tKdUKŏ_aw7J/pA k^=p[\#J>z#9Zwe{ub MWA8 4~os@B:.ʕZN#qׁѣ }W|\)Y'_cT#-]dr%G4̄Jd a\uxU/q55-BڕUxNzpq@T3AS r(P uTCsv-!>x+]bcrU7iW^@/ISFX$ПKՙ0.k x;ޜl[W\*sB7"q֓~!,i CgB1{ v+OKAxL)Ǜ -K`=|t!NvϘqprmں9AWn'(}[Pۄw owAغ7eLˑt>y?ۖ`.Ȭw<љfHlؾNu7cJ֦LA3渾xӿ},7oeGI, '^L^h>҅60k4 tNΗ`ic#'ھSZfݻyzEMc0$Jp]޵m%З{]0$[,Mr X :p+s뚵PT] فz=:H[K@^pktbH#*{t!J\|Wl"Y)ɸ6!Z)u<%P-"H%":3MΜg%2<*j-L>n`0Uj/Ɖ;̄UsۍagEN{ĞQb 'PHQ15"m'!ѠJZw?8Rg{x8RJxE0#_tpéu] %'osQ`zظ29E8HR\ \4Qs5Y@(S(S{+'^ Kہ./QЈ\9<%'u_qB7^S1!ȑgBfxkvFR8=dW{n;4&|Hʏ&wȉ8r_K)AislpE &0z⋖SB|R wJZv6t6 C9v &͖{ev'#j]v!$gm`TIv9xU tlbLsv4_m,=VNѫi)۷ю$p>Vj:Qû/ 7Vm '߾vo>9'&dDKdr;M'SX;Ww[OK$*JMAva0/fp;FIf& _8Fz}BY/Ӣ-8PO?6L!O7Fj}HXjbL87xbV;:Fa: o\8%GKDR\`٧< ܯпw187M>Na"z7}}PŚ&,Ƌ,ZU{"~߲ ̄=ֶAާ%8å~?{__^Sⴷ2F%$wiak #NZ&c.h{5]|g%e׋/e?d WUOKp4ǜreD?o8{uME"W=^F R=hgZ!k!Ѡ(9(+kNP Cnٺ^Rr"*rF r *MY rA+) 1]yFBMzc.ȭW鋰%<N)5<s' lA^8Rp:L<` 0d {lE4bJ͹ޣM'(Se_׬AlvlQ=Tgt&Gt-y"temi v, OGƼ5w#]w6#ToS~~0krPog4̆#_E"㬦ì¡tc8t A18:8lR444B2͏4^>ÜZ!8:6>.Ȧ{|0Isu|^#GwGNN8$SҮ9 cgp09;SڦJn N`1\ɐ#j L `28)$E.àktI5KiJ|JyKh魲2KdS4RV GK%2Y͌XB>Z-(Cи9`t|`ik($ &F0ţKy( 4\焙finp7n; մXx'b! hQB3iILaa:##":Y"f^R إ<7a0I-'BHy3A3er іU80йt(2P.9ɴ Ѡ[ 5maXoHb ~|8 \cyU {Qt@tEN~_D`P,pT+gM ZTt`HD<%DJ| SϛR'~a]>HuQrlPWga𬡛?5+(㧏ߑwo?=>)nO_""G$[?!/7=pǓ&#OU[3eWhҏ!\\C[,S] ASJto= R R..@f6h H(1qDF4S}\7No&3/aZ8o¯86xĆq#`2y~Vhip 3.|Gڞ;VR 5,]H|tbpHEh5Jޣz+R;WP~bQJ7]ɜ}|a7eh d%q3,F:qrq<3[>{I6'0q%Jaz0<&"OX$ fiӷwŁd+ Oj,Jp!UK3u :ݯu6û%χ藪2FOG=Hm4\,,>KHc3 k`CH%Kt]{o o>1LY/++Mi41Z5g5.ms Nn<3x %o ED)MӍ'dS4RYUE_rRWW+-+V\xnsZJn=B뻮!9RV8f1FG\\ˡCwrWcUs>Ixy+Ic-&׊ >":oe#6gYy /8{[8=f?fTeU+ӬN 5UX_='͹ɿoU'Tĺ&_xHj?ŵ~.S_ &i\n\yLje+:$\#e*q*Z̓쯕aʭfWhSR9^8 qvrVk!{5|XmB!s,jk:xb}6n|Z022H\}2Tvsf/lpaǏ$'dz]~5ہ^J北Ps(kTluRql$']dBNJ.:MsDkV>[nkCM"tJK&g+$jU t-9Bw,aV9i,-#}ݾB-ijբe8~r MH;b?(Cp!jsĢ5$e|~!R|c$9f{Zp?nlÍ{1S8l}z^3n:GM۱vS8$i=TVd/Ayuh#du@ E8%GKa"i/VփUכ)tH'}NrF_ {RPgDk ɭE:k _VAr((0R4jkL2$-$*z'-CtY$!FGO݈Dl$>t{M.K&u;*X"Lxb6 xL7ֹ9Y`VXPDE2|aS@SZx4qw!Wf, 7 A>gM烙f~Paz[|kz 凯ֶ쥇rԖ~nM1!lN/;K8bNT Nfdqr&rI"2~qAzQDhxFL^W>EOYryXJQ!sNA1[&jRyVK9 V:c(E )DLS!)#a['2tp66CO0AXN34 i/J2F j G<#1"\/ȫFBzC9X 5ݤˢ+6Yo1~'d1ܞ<3g2юL%3Ln%hkD,~rA># S)'cèw:pǀtD fxkt1 *U,e}tcah -8yo{kaD*B r]ksGv+(~ʦq4ؚ$)kLT=.1rU}vs[Лxo0x,p>lz{0sٯHퟝ*'ufϣN.)%Zl&*2Xe96cX8W#ede!򌔐+ xVE{CCҿ Śg|~: 8luC%à> Fcp4ک:Bs|HXL3Ɲ()FNE`Drz`p@p*( }(Lg.sysAE[߰v5tjA q}dr^"z5VHJ|ONը䩔%)LuC%dVmXAS_AtR7QmBܺ [X,0p`{(@%gt_%S.~_s0aIajPТ0A!,.\¾lR0~A3&w kp6 B ¯><%p.7S"%zJrV{\@ S\U`8\TAZ39j(p3e:j2H 2R3<T?P^FyA%J1s\iYW쨹bü0T>upoC?;nAꗿ8?8e?Y[zuW_럟;i>%ewSnʗhx{ꮤGrӠO;!gETRXFH,񄳛-&]Cf7S8T{vn梇ySDgnnE1b`ul1dvR lvn梇yߧ$& {kޕ"4"+5UScjx[c:Sӭ'Q -뷛~{bw>6u+SG4ސ#}ΗlC1 vwI2߈rD=}SlKáG4G7aq[ t჉ 1WFտbްSFN@a$B (LP<ڕtb[X{N ,%d#N0 x`kVtv W0I Ul^Ac4#[~݇T7-Cgq.֣y;`Ch %%O_*1uE1 Ayw#ACޔ> 1}zhJ[P-BOi1QCND$RR#[R(1C6om;j%IknЪWūzD/U1'/1K*W>QL_"2jYS Gԩ HPJim$:]ppʯ~:8PpE;ʢpM,K*BlV7-&o躅xH9o#NdU:*! 0. GC6Y\qE~փykY<% q<&$S($ϽCE`03{b耴Wa 4 21W v![ [pv]O;緄ÎS`O}%YV,f#Q9WJ2*(9eUA58C\J*;1YszY- @9 _bʜA^APPYrr H!B,d)F9geCx޾~jviX:Va{s zMUN*l\\$,Y!pn`P@w?3Mz+aJWt@Qu/[-$Pfj`’ͶõƐmg4PK6;j%]36} ?Qf7[ L7n[42uE9.jӧPVK;^}oMgv~]P.vV\W_tɟQ&v{Czɻ|zKve]E]yW_M̟b۟ߩ׷w`'jJ(Q*7OWZQ"oBKJfސ2ADfqA`4(+2+J $'R4gpha5GqMe՗)9>?/50 +Gϰnʔj qbP i(W2mdkR=̒L?w( jITb(pռkįi ^<2Ǧgsnv]P|X$]?V6#!c^}VN<·yHvcμii0"G (ψdO*}j$n )԰3j*2:^!(2BK+1WXj"\ejg9~V{j.Ra~餘XP*-!P q14-/ 278$bbF)ͫr1񍯁.QˆOR2#2'dd Cj8a 3CÊҪ\Hr#S'AB#6}47k,JAlBQ^X"}_VPMv)zE\Qb!};2G UH*QXi!J,*T8e"UŊ2̋pSLzˀuxmhQ88LG<<::`l=.O.߿Pvd%.-;z`$/>ͷ./ǚ0'51u|.7-:No@D C]4tk7>Rul kЏ]xbnV6o [ln/W!>bV$]K mȏA(9qI ' "L9 Qw幚p.>@:fχ ͋"4ØIEz uNEl!UWj: mH){ ݤ9 F=/A:-祈ԈGΫ IC%x"sR&ey=S 2L^~ g'5<̙ċu8z\z\`r6Tf9 ^e/jGc7=E8 &cpfT6N:! (B̈́(_" :GFa%IY 5LX_{F';4/r^#P8(w*1Ϲs+˧sR$3:* dȆݑ!(4H_} Dl(`p-꽄R i90 A(%J)Εӗ9Lz|rbặ_$sT.jÀSDkC(>>#B.28s>8{rRA4|Hs$D]UCȝY]N%WšXNCbP!!ISx.4E 9`B夙گJg@0I(̺7XBIuOmcJǞw $ў ]B3},H Yx5XVq9["[s n^.fY~Tzb`Y[Y:[gElIm$<fk<=(=v N8'j;T=>'<yqf9B8(#sWT6J(r7CԽSɼa:]ia!o-[_g/LWCV1?mlwe^~>"kWr=av׳;Z|ͬ~OI  7[#{y'Yg1F;f #djm:`yv`m1r t̏D5yG'Z6 ЅIᧃsy،ct>QWAَsvryD!gp`۽.fGXx`Sv(KԌ%FWS!FL)2 r㋆[lji7%TFs_ !Bl%K#܏G79Ef-{o~K5 Sb1(twUc2 ] ]{SDXPj ¤0l _sfxY?L8HK☋L4z8Fb A[pMGJL*-hB X9^6N"XPP@e)=lFlfD'pDN/Acp? I`,O47JrMwX+Tn[Ǯ:,lCϮD-!`A2HLΙ { Q:+.DOBGiRO 089T}l:x+8L9xDBkLB !&H{CAjuM"$"DBt)L]yV?}/_]{f&KW<Yo[5ku2_ݑbWʋxu,O+$ E:"$PjXߤjNXi^]K.Wգ]SWW؇Uy5[>l6mnJ>6\vQ&,>ӟ<_}Kjvv1t (uu6)V2/7eHW}euw?\n{?Lje:D$؃uƛ( 44_#/#KPoNa]ӗVbK$.ԴyC,&)!bθVLevrua:L'y,2!R D Vͧtu3n?}1ym_RMBR&$/:l$)-ƛA%$]j[˾~7?uѤȒΜjOߞ֮:go"2KgWjٜ.Ce`HEM$D-)H 0%8ǖ,aj WHEZU\3KPU)dJ*Rc 9'9,*? 6l(N˵WVFH$ ~zvuYtITv5LOOwOOڶio0z4CvH3`4&Qj~2mz?W3GF)1z04˜LxȏrGHT:r]yLsHZ3PWR2 پ {~ua|𴌋[v=.{gSPC3FmZH#tf :rկ<TXcT[(}gQxv:CvO W=t 4:q΋g6G'Ur΃ )(beXQV~銟 ͊ԴnA޻L,ݼ4=c tO[oö8ѭZ`I#V?͛±>u0B6 ۘF{f΃43T3IE~䘙v.lR OBOtz~9z4JOY*d߆]$T<: BX0d]> y!6O_lxX/pWZ1d#KȒfYR3[Ȅyɞe" VCyILj%}VYr#47%G,bJݗ'd44[NهX1h; ZY2rMr9xƘ+Hiqdk dsFemоi3:mvܺ3ǘ ٳ_1][Eg!%&2t0k&x:(n$C;&5!&Ht2^$f5l T5u.'?T%s*r:ozIݤH߻XIf۽A](Xn=t ;=]߷ӯb9Qݹ 0U*4c[ii{aiU_uw\y4vyտZcs;݃MW8U]??[`Q]=oσO?vp~rOmg~;yOn>6P:>*2p˓PA̵CH+(X:\?QG:|C}D,<'p׼FSK_jVMhy/GǗ'o?\h} ;6\-Vbu|89=X+qpuϯ:|{g_N8r3>~8?8ջ}pΏ]PStiv >?;ӛO=M:1O˧zrk gc5rp7ٜ[[^CLroyޯ[0.BԝC8xz n]!M #$:^DãL<{ 7,i- G^]Z#y{(b1%>߆RLKzD@j M UnIO#7ݖn O٫#hu8; 9o` ZYI[w8|$0^}h6.n}̭YN7LA{S{5޶-x6a9M?aNRZ#hR w-(zZf;{ex3{WWMe"ͷd!nF GUzd^,]S̜ dE+4#j9qJJtb! UTsX^_gϝo_i]ou.*yё7\:E"}<#::+Pmn z68(Q,=PA[ > ,+!+gSN4Dc'P>rL\%}$ $ r0w=" Q"E|)4pE$B(1=Wc _9CWLk;ɏm>p0L Xp1NO{q+ vz۬/*XCb.Ȣ8`"f17NFcbB&%d,E=Gpe"1Ą39^\!AJڽJ-Ʒ /v؉wK-FFPٰjp7@oJOrky653}\Y ["ta؎HrPmY=ebQ'BybbRbvɜOZFټD+ XZP+x҂SNĚ[}{ty1ƜdXL\yU趏7&\_gݰ)[CD}@ W9i4F<|^~Skw'5Ei巵 UvVᐔCUn*eΩ-)Es+p*&^;'1T('xyT^N(vFNa݃+UX\Aοrg$sw< ε0/3(Ĥ[E0q9b>AD9LP_DG4| R%[$2w 6JK5-q&ar٣jg<@ ȓπF7@Hq$^D/ opM} 1jZ©n?֣HF[|0As[6d3M͌ɮ*F$Mq^hFhԮMes} \}*Zmiʇt,6NyV+C\ڄR&$gwrII8W9SRk-#j^ fC/a !ـr%SP띕0јI(CXkHhh2 -:o6@aCX"taE,oc*Kb|Cٖәl+h5Z**@'e׬p>-aʤǵs(y, ;{(0ZrZ6=E jHorԐ"\a(%7S956uAY_^%d2M\iP.[bPeRݗGLK]N&ca0ZQyyqa"4-r7Z(8buKsB.7 !d ތX&vP Z X(ʳЦ p\k=2C ?9Ȅ5-m2 y>/O"'2V$tUJ|]34Iy:5eHOY98:Z<tfw [٥&Ex|͟n2 W K2r|xVӌgќ\ nx|av:Q ,o(gx3dMR56~Ҝ7eIG d3)z.Pahs y{j 5i+qFEw30k;ANlHl͞1_[GK>( 헵!6UPSy~}98AhwXEy$Ŗ m<^$_vȉnqcSGxsc SdsސG.:T*b1cŃy<`4HoMB,a@ vDkbdʀ'в8XHih0҈j i 1Ah1]4_S;: K72Y2^8Tg,l}^XWl}f䄣~۔%9yrnrtD[&*+! }YGB!?YLYtO+-\唐Rs1Sl>Qzҏ0)8 aED,J6Ub !\-U=%IYq>˵}62=Ap Ⱦܸ\{|eW=XZU^S]ĵs7G:mz7ek635*GD:&ڊ+q<*;ղ //)BZ@87ڥ%4׊j7%MTD+E`┅r0OWy'ȉ'JMYTGVD+_RFa6͂[R^Ld[ʿ#KDɋ ;@ +i:PT޿' mrGa$os@ޔ ($cstYlhckSGiۯJ^Kbl?<ƳJT"OiVQ ZQL`' yt ;#Fw^*.Eiq4Wa߰Q+@弩26p{\5Hj|BX|u8~3>6fΧ+^G9>@33B "0+.=` !+)}^)W&V}gLY<(<*xP fK0 3u/AQaW|/*6yd| iBZ!zim~Ǩ($aQ"JqgCx/6YR[Sx 嵗CQSٞ4 AVF@kQGyy*uw#;UMo^S ;`^7IEӇcw^ȫYÇU#S9&^tίr | ҝKd Y (cv9mcoGf+I㫱I XwsSaoY:֔^}%gA$'GsIl#+Wu.IW?Dt{:;8tts|1遧cz(# ڝ0A0^{VOA ]r#\•rW hjWz^DMtkZq8;ӕs'V6 _jF=i㲤SgV9hH\_MLde>j!h\t4.8с{c@rD뎈qG"bvekw1rNA;n[qFLǘG(LjED F+41P(ĸ "ZOJSmk!~lCU1TC-=}]sʨ\FAրP 4B1WS8\"MX,bC֙{v1$# WHhDL3B2N!S#;y)e3By<"7ymaSg`Xnuhib^={VE}P BA??~ |05O.FXLM=S#uS&Wo25tdo d)t\K(͏ǜ# -?sLwo5|tp3ԅ,4ɒakeGaE)]M ʒ;J..a eXےTf CXwa K<5T^<c\zHL'`ɹɒ?N9)YÙcn°4voi`X{rh}Nm;֕K ő(/޽zvUmhL$tHUy 7㻥ѐF1 R}BBHZޟlɹ(*C0Y?u>%dr?TxM/NWEq蜎guehvaRJg] @6)K`1#mc@/זP-~Dеa~1 ꎶоr~>FF)EbD! BZ1,q9VamͧYhp?{݋'Z<^t9OQ5k0UX?m=0pTuj4ˊ.̪~f3x܌Cgm?(2ꧦ&}'HyycE! eKnD( 5 )d"eE-́dڟ+.7 ɶn\0ȆiM&֖?m@^9N;'U@m|@V#d)CmjӞTuHvчe 9:e;Ei\1ٻfxsBdnE\}7AdU_2'e\![ eV \3 =} tF7MȂ~8‰itp#eW'&I%>~O 8OzOH[Tz<:#{/s˝?9xGXe - 6l9"";%,K>IZ@)\;wz wFmFS8.CpJԙdQɷ韼y_|Wy0? :V^{ڽM_~oW/~̓::m ?|$yާ^ ԑC5s=|s)R@7&]YAd[Sk?Ky^E_rroH2\^杝]2&]i4,x=kRT̈́ͮȖ _ޙD)g =4L6H,7Ep]D[.Sܻ I_.ާќ5J9v,').7o01wѬ dއ4؃amcvn[NU7$G2zh`ыAlW=,5`-fUcA|d>]f\Bs:x8_??-&5AoYn[ypzo*oU2d#EGW;T_xz<}swwkB] V.+!OduzW;uǟ2#vNl97w'nu&Xc5LM߾m*\mJ}z7n: W|&t笆P~+ؘ;2QXӰVe+*F?ͤP+ 2 -".2X6E^&<| )jWAt) gkFԓs@8l^a0oSst ^65sĭ~?[F>_8(Z߫P"iߧ=O>qI~r'gb#eh5j7&RjZ`b#bС(Kc;Jb ҔhE&e\R2c5Oh1777^mX=/vu Q|s8*e4vbYEbØ[ iE(V!ƀs  J*B\O5jD,ۉsxMrMW7?{ב ՗jz,g$ ۛ *q" O 9΅?ړ_Vr- hmE۶>Yw4fe1{r/[^e˽\]4fkOcjHFEw$fS5py V@%9)gmbau6f.4ȘlYt4EgjڝqV U@cDmIDCLeEMb.j9EE Ջ15Ql~ ^ɷGUC- awhM5QK睿˯CN(qX%T`žV).E#|U] KY}nt˚go>\ߜk'JOOKs]i.@mF>g}a ->?~usu~tbVHtBZwaHL2don͍՞k|t96;*>s> ܵs[s*1zv(6K&]{םjU)x*-M٫6A&"-ʺة i9E%AE(ȓoK5jTH8OV4Ô"Qt颱"PTc Ɉq®km s{WNdnAmֳDXwkeBBNϙRL,f${7*TBjk$yXkV\ԵPhW̻nɋҶbk[pw+v]~{{~5$YhXz(<~2$Aa$dc>DK nB/D*?"0rβ ~-y̏ެx}ZQA[;qo ,2i]ai^~珽M@6yUrg{5u`[H9;;b(cχ9Bu.]X@GNtTITF9-s,M>/nQ=^)sr20!o=*wˆ5N@YGer|w_*v_uՋت.L~Dpcr^InO5P5Cy/Jڞ~ ZYwen,~wCpd:]7 Õ ~]s,u4|*_wecz 펺#x}Pl2WisӮy6Br & ]rSv +HGtkVN#; )-ْȅ`\ vm.oX{sJifϋoCGNL˅L} *1g % 5ZPmLXWQ:Q'L6c jp5ی%Y]B$s"+əTV!rL'hbjs^D9Rz,8;ejNN0.-hR׀C O6'SCMA8X5` eA9I[ɑTEJ>RcmBX6ߵ , e;͆$=nZN3?K ͌,(s WuNӣ M^Ĭ՜VE\ tN|i%&v, ~/3k'hmh{L>\ Ռr+oQ{(4b0mdeN gE)&oRZ7B!>Fw~c% kR.:qP~#m5؁l\_~fRQZd{~CJE˙W89MG|{]B+E#u{ߍr=>o?w-ZwڃөW(Xl ag IDeR2 ]( VK |V2Au9/moQ[uUOl\_,V 1TJN` l+ ل* Ly1%Is媡 \5tY[Qdm㘽f՘1T0(,B)YI!7[~ܴ~Wu,[׼eޜo~ ^V)%^مҹq4{mE-4e֘ NeUJ.[i(՗ODiBEnl\kUGm_uKbXDmv†'ZifmxоT@ӂp0}%S*B"$2ć" "[0rb1f:K@hgϨÞdtZ=D8Y^Ah˹BQQrP72R~DQ!9jff]NwDc_=X5x0mqVŜ.#^QGp\3GgZs.88j󕓸{l_-Ӑ7#:n䀼kH7k!w?Yrͤ=8ѭg?/FX{#o*uZOVm95artGbz ^~<]Nu\~]~OknEcBxt*qWB8 @N&su)w~CK"rnPjC|e)r_lZykN?7>ݴ&g34@Ti/r gĸ)g+╭o9d"WM<5I ces_nЧ7_-Rigm!a]L(8uɗs>{@>_jG F yX`'^A5'eH._4I޻Slp~mz/eOШcfSsjlڧ.\H6!h^߾ޔ)P uS{ZX'7qԐQ>lwQLl;=FAhHߋsNu'xS1ֺW_?pכÅr7[%1xv6Y^qՋ?}l?_ןd*cg HN_"AwK!JC) dG9씢׊_=j\f5?f|ޜ݈/۬$k-\G;ѽ5@ g\3c &Z]7J5C.aϪ AOfc|捳 y8{f5R޼&+4&i}Ь/\nÇl [7^NKlM{ؾZ9sH>`N'xW9LjbιNU޹tX1fotu5rvuHhxJ0P0'`A1Xtt9XZMæ{'m?l(˱cN}QD/GX"fGu xQ.VEICw87c@g4a~FcF}z;W~Tp<нFDCh'>. |6wCZg}1`:6mN3v mH;#ׇ.Oq.Zcv+`>2&q.O~ [#y1 hTGf\U}/»2Qcn,_?nv| 9ދlS6lsjlc[fcA6vAQBT;,(JIF^(A)I: g)SIFTЦ856i\Jc˾|12JlXo8 &)XI9jҎrvFefJ9ǤLMb]sQ͓/sCInۇ7uvu~/sʗ?ݐ[z=.6o|D]JJ9dkEOlIB5Yup \@a'Aaֈ|wbCvmI6; $R4BQOE3m9p۱*STtAWb3Ї r)Ued61mD*M/NhI"{ - %;#m$LM @;7tC ULkQKdZ=wi&^xs4mBg=ߞFfA\csG&f_s*A/p6kiމ%WRYt"z̚՛["TGkTF5p åB^Xos_um tbnVŅ\zM@aZu85Q[¢/oo߾V uP$'lWJV6hg-}t҂;Awg5Q4~z]#*ߖXoOEw޵,ٿ"b.v?_ 7̧ H6ikFIc-%RSd!10Ǣ9UMf?]% E_FEyT40D$d+lW!U_"[WV˕! U"e>ϲA䰱%cާ Xf"S*PC W{5VŲo$_Awvia$yz|i5Ѝʏ=B̾rYN<#5Aa'ru H*gyҟ9SS 3[WCk8+'J;x˟T:J%X*JaW D:@ES<أg۰׭bw՝4 Y|(_| d4d6.dS`;BϨT3y pj|X\ gXծn=sDY~pqƌɚ\cS̩*& Mp:Phc%3s'(ٿn[|?a6=Hh?n650Mv:q蔫[ DbN[ŎHK! 9hS܉0++ ɋ A1,~GAkˡg3&3ǔ_+1IXO"+" :X'yP b 3!)8(,fW[' 0 þfT 4ה%Qh;~€8a@O#<Ca3medX'j#.]FA?Kb5-nƈ=?fV\h@NOݗp~7Gw/T{ 'v ; m8rFL1Vh͒Ix\q$ 10n!tKcᆠ3q-f !4$ bkƨ4ICpTJ fFؼdq]vv ]քڶ"GL$U$OU$Dǘ449Jp- Q;!9Ƽ(IQHJ 66 Ӕ8MBL0LHE@'%Dop<|761<;n WXI"S[e0 tF/Ԯqܿ)5rce=__qH'4N1c+n-۞n3X-|5% ~Jΐ9K,;]b69bjg;cG"јS7ggG\ɐ0ȶK͵*~۝.% ؄ԘV%2q]cmTBI"NԡV5\zG<C„s$2%V )L&bmV|cory0c$SA7O,Q dvYhNiHko) MVRPAY2!u-3*sxul|L(e0%'7zp?r5b <1E7lym$ql<Ĺ6]J ,j_`(05Lyi8^YR;uZ`o̼X╔F8`4 LT`҄0&+0d4+/69HJfdXF8A$Y c'D$ܘ w3EU F0 8 H'= Cxg*|w>hTL\;1o-`+k/do(_F첂H9-{-bHIiξaÆr(@M^Jr9\V^ ||o>#ss8͟y6g drP`DoRo)y&ܤAGHuJH"o 7pd5<* kvԾ0)C!ٻj d*=5:tk ;-pl!=W5/BEVlw)1z.iR$F~!g:L8E05NDJ2$p6/W%!f-|ui7⏦xj?[+]ρ}0)HDRfD8X3?(Jd2TLjc0EHP:+]#(SO D1~FEfge\Kk_s za"hEsRu0_ctnPQ=ߥsft}u3_6@š4msfXtCĜK0G)d)- OOV'^v@e5u`I,T}re_CtӅt4u<}e'l~[ETJ3{>QրB^12& zHpҢJ:Q^^Bo8+8'"уsopiA'f|T̗˅z$p TOCmKiN_KCdpA1ir,NfQuJ[FPЗ9ݍw04pE8iIOs` "XĄdAԟ* Y;ӏwL%ku),ߙx0/y~%/p;d+QJzz0 0k:`]"<%mZqy$F VP_Ii[=/z\qR;ă_8!^1 Vj_b޵Td[7t\3753덤_wLf7z@rQJTFQs(I h- @H\{dt쏸?.~[)ʼnÇbȕ!eXVV,\Q̥OtMF) sp5D 1& hmՌ`mZ3>ED\eJDD,!%6:T]'Is}J3v peD]fG)T-!&X 86jApuI9pTEB83Ҳ03[ l'Zԕ"/e1nhP0  f FqSΨH!V6lٟrb B?|]>8Nvhy؟:ǣI >I 7VeIuJFs<#;޿<Ί9uG/LwN|#0]:OirOK2/:r}?qutCUw˂5{7\wB&=fsϽs:"[{؜e7} N V$/~`(DsI2X&n~F@fcc^26d2(S6N @II3هr RΩ_}( tTd6/Ae ^}Rcnz.7w"@MG23 PUxĿw!PGSULNzŋ0ǯ}i -_` k}VsjA -~*S1Néi3r i \%yc4M:~iƀ;N'~qvE_&'"@Ґ-#{i>.+9G/YvpG|5I4 ~-1tK36Q]{Ss۷_;0u].@Gow@MޕEpqhx2_ W j=ы`uQb j5d4gaL?C7)rC?xbr1}0Xg;7^+s":ͬ> 5Y:8w>7__ណsAJ&LxsV.y}V?]zt,(Ք2M8(zs/km §] l1 |ο&MQ'm tZ_f3x!VΘ7!.`nAp6_ T?#pvUK]<_kÛÃvzfyd5e؄6'siJdi!YIX(;d$$a&[A!c?^eD!W/m]_}24.{ Y?F0} Rn~҅k)V-3l8̣(#M`+ƊSobc{bƤ0H;<._ySʝn'Z1;T@Նz-14i a:#Y4NUnÖ/q閧, *՗h HdJ^_\""8~{<\#&KWTQ]eo۽wwwws}+'!KV]UF+%daLene&F[a,;5n(6 6 x \;@Z(bpdYli4?YI` x sNaR ˊ9ӑu"8 -wPUBa- zV;+e$ 4-j8F))1X qEUVOurֵ߀1&sGDE9'4B1eDb T ]ֈ*Ә֮,sac+JH Zꨫ N)#NP bFDxmA J+cX:yLyLp;"M1D%]" "3X)yI:Üٷ>u~c> nnO}>(Y`~C@% ].FxtiE1c(@j C k4 Ѹ fl򊣵, [~Q6+t=%V4bpj) /*pb)VrD^Xzʳ$\&/$=f&{\H%z,{[u&)1$XdDw㦒PTWӼsc2)hgˉYgM !ddrqEKI6r z0BTAP "x-T콬3A$N( 嬲F)gB8S0y mNb 㐨G5)A{BWÇ:]6!~ke,,%{$/T^ƑJE 1xJL8Xne50~GX؂{ GcrN1C)4s }(f!2eN:*ke18YBcmk!b ڈ)^i#`Idu,`=)oz|wY%0oSA4q=r%vͿӗܼU^_ŠrnU}&R߶IkfL4CS<6x*rrūec P7zC 9'A>D^A dM廬?8[I#q!n JyJ!}k݄ A{o&2krwSp6O,Z83dK4縖+^pxb{?oqkxwJVjWN-̻%YQ;ӤVy]Nc*֜l/UygenZEl2o2E+8m)OJH1%.ªis|lͳJzwX PJü A017a hI|X>Vrlǜ.Jlݹ]4}疃k4`mdۘwf %ycn#֗1ܚKd9*5sTiчܥ9޺H'8ʉBs׋QlRyۘ7^ tE㰅v+7dfh_8&˕C?M'aX;@)!i"*;ޖGܚY؍n]٭uD<kT9lJXP*pe8mpRG,(0:0N7{IkNp7` K2š[5а(eo\xݬ;sk~abە;ؔyZ4mRXbÁ{R.pO9pO#p#wjºͳ+%?B–O"$zI ;1D"b5bG`,8#[xRaDq^΢ƜYjFOw)4 ]7Ψmatspbg̿ǧC$E?+jU5tԓ{,>j)0 ADaSrxB,i8SO'x,po4}%YO")aJyyCL$Er':]Tɮ{f{"WȕmнBǒ *_UEt *M.\‚ YgwB׋GaydF4f5 oT#U\}B}j 4>:w} EaYnwA]{ҳ *AUÔ^>}8XO)Z"MG>$oRVkAfH\0S1ML EccFͰZ{$v(SBhJ 2"DQCdqǑ@AY@ QeXK@Zi$s[E *#M\ViZ `VN$p%?ΩR*kjJclr}DX6ldcJ6|xmY@߱v~gGOFxE/+9lH^N4ƨp޲?ȺpZ#h}2;RHS,O3Nl\~{^IxHm;;R~`)7hC+v ͙w}r~!; @[K~?c;gJ&]"4$amd1~[ʗBG 3RN|2@4<)xS6M Zh`0W=-e ZI1DW^Kg ð 4]O?G#|WvBk q]%ʡ4Φ eX+ 8`/֟%Mxضw/g\~z[o1a ͠Qm)^E-:@5/>]^(xI'!JeCOAӆl;`%M>?$r@i4m3vwH:1#yneSzy"yjn~8KTEhB77{j'grz@q൏׭ Îe=r hs{.x. ^)(Ԝs^~v-V1U,k*`":AKKda􋥬,"e;( Ùg/L {jzR+vXw2UwſǸ#SzVYv@wX--`:uMtp^<]L?X^xݹ՛WuueJ(:rry6s;~2VC_2BhY\́+r'6t?|sZS/׳'!a #K9὾qܽ[x،f5' l?DtaI,J 1\+TZhfl6YQGLTzrHX0 x(gK*! 6&dZ%m"轋!t|f$AR]PsNglJFghuL|maB`͘ m׳"y}ֳgYw#)PRӭvx r5) ,$sOACGSLvO@p#$'v3c#fdkh4ҮV' cc )(`/^dz\e>j yp3(XAXxGGhx m6>L?;{Md~uPETt?,JզZ>^v]'irZWwG ٓr- R%FJYYQcnsXr;LPiݦMEJY yjL8Veg@UsՏdL +/W壘TsW"mɇR bX4C+pFHR1S|cb0Zjﭥ{XIZ}ݳ|_ ` `@a(?rZ4l=]ff [|yo.+@X!S7L x53hҦQH?/v_6Aml2p O>UB F~`pdJ! wJO}xmMavuI/7k%kwyP NFJS¨" 5W>Z*?˵Bl.[ݺ$枈f<}4IZשܯ 뱬^xEGWI<9 jD-D=)$7JW_ˍy%#oɮm>izA|ɚ-< H9MF*Rν#;>#)b}g︭bs1UB8V))Bhf2Dȵ{_U&1;f쁝;4z֚<m!v(7چH֝fZd htVGm[ؾ Pxe9Ԑzewd-/SŴ4t(f-E-X'uX&8_.~X^glye t>zʳ5̹͵*Zc"o5͈HR.$%2IɻuMFƂaހ2DAK.r%Y)*_𓷢"/+.L-]-?s_/O+K\"P7Lw#<\7ߴLX1v,f@w=ZԳ*0 IQ iʏl)tI"1OPt'@V%g9U>3T*Q' h (=eÎ 8(b۫UR\>_^f z&$!: 9/iVE"J"egHIdZuɞmU@_cƻADb!0>O9#w ,eZ`kNg8&g}'nH"0jGS$T#0!KҎ, $J$*#eOZ碥5D-:# ^@b0 #Y7+a*-YE{eÜ%0/iJv>:KA%IX](r퐲}縲V@(ɯQ3aQBA3ã~Q @-},V+ɽ^Y琱v&K L}dڜr :9_Yh dLf" i#(i2L6Vi0S$STZD M ׊BVVi@EkD UkHL+6)Jo)\ _PvֱF*Vy)3M"9'm@GnK`:[ykhewN  $E4S)CoY!jyfJFN2~Px4hW䟭L僌X52jɇs‘{mqwwR}Rr.P ~J PJRJMrU&f|lyI3V7ZjwA-?Dzew Ge8-Īn6.@f>ݣ|R4 }Z@0p9?) ْ#JqB ?h`LD5),D= k),eO4c;?,=#Emlb` "޽Z[JkÉ:˳8`y$;τs[K=YX bmSl}%Zg=w#>>xW8Z%Ňij]Oo|z'uuoKХ.w9@ /!05怹Ǣo+!I0@xeH)-RfPEN.B(ÍJ6:5s*A]"+9k aC! PJdB+EL&XmS>k%Th $scꊇ -YCAv-dTKeh 2=곿iHwdZ#,L3SMZV&;Fiqwz8 *mee1ˌn[ [Ha(xzD1*4keROa#1246J__JL sj壅׹XyH%ÑJX*|EoHyX ec-|A"- W7D[ :&uCFWJvN_ !]sAC|;Cc)K7}->wť% 1*nQT]:jq㢓+otxN#CRXj)J H{{7g?U+h8JZŇ俶7j=΅'i;ڻYe׎vw1jշғ+p 'V! ΄lN؛?xw/W&I1s{@s Ğ;2zkdZfv9o^wrm pb!>~Uo(T_ 6DbH fP Y=*B%v^ȈICH -&`R;NS\UUG5# btmƢBO!zƔdJ6$A,KTV2] 7R I,ҺTx%-ZxZvNxpӹ@?D;f.&wWZ^x|G>ͫO溺\2?uC\UGL_VKFhYGŔtiGږ %1g -?GɥT'g#._q2bew.vl۞ CZhqyO|p1rm52N Jk@K:%DhVMdpO?{0x@eJ%z7%k2| NX7Vk,R4"II  B3cs(fIAX3ȚBd͘kg"nҜ5 -* ѡ iҋm0wA&+j[bCqu ܌5t_L* 4<#5[I,+KT˗jO.)kk蹺xh"t ڰCƖ؎t6ī'9}x]s!C68_#]簾LgSS/ +v5p1UB﯇[Yd )JGoϿ!n<ةj\')޸#C/0g$xBWF'~1:jhQL&rUOxiZَJ޵q+ٿ2O{72EҀ/bdt/ >=h+iEVKfF2Sb,wQ/&&}^g9ZCvw(~ZAW 9}]P!m8;^.aZ 8=ѵD"⨫-Qg 6K3Q&pExjtxO3,bt U!H 82ebԥZX-d4=n$u%vbsvw؍IZpf YuSei][NಟnIeK}M]J|pgYbda'vXg3RU+IS 嬒Q&wnVR-Hx{qǷ,+)mҋI3o~ro2~rͳN<՗iߠ ÎdØcx}XW ZrN*+,|=ߠvFc}Z&2*ଚw@VUp)6hB]>q!+${rt+9?rwR؇%UJ 94Z߲tuoy.2Kfi|#JܒO%ls?f*7!_U}ZBn/Gݦva_~2o$eՕ, ,R A+T(ɩlg꺚+Ig9O:5l Q#w+ꇞg-,@yLGyBt274YR9L}NK%\17SWWSb+/7Zwi 0= o.nfւvf.hjYueka :0XsC[vT=G^lyl˖/ 9K1e1`4;F-vӬL3ZRF N=eKᜂ5f,ٰԢ,9}RL٠;ף)E*SF2~.},"ȫo?z#q`Bq78IA&*X_xᄁȠ) \_i86Qʧ:?dN5~%FՀ "/ M Kt:etRumbPRcW||\vRf|`7(x`Њq=^Yʙ6\Yz?d,$rmx"ROR5RE&H<k5fa}=gD3sY9HAb(dNUA*'~:FJ R̄؅o8|U2uuoѥ뮪fp 3Dü20!č|qi.ϟ OYn))|Wb* %\<}l̿F+?*5# Šj0)wU's-4?!/y!<́/$[HQwa@32g;7ôQlіY_€Ala/Y^8h4ICṡP R*KLs~Zy;U\&~\@惛:|WD薮m.~SVuٲ덂UqVHJTl M\΢h-#ybt>},q.`:nJB퇬\(4xI}n5jq_X%1N'9$ *9om{HB g]/1r;E FO sxsQE.wva/8;?\BiO2N眷qAz#FV軏Ev#TT+ucԗPet&:,a&BBut?ۙz@.U fJ5i&N&"_pF9^ Lj"'p-nB2aMXP2#h#זnBhKzIQ䓋kӴFB*!hIF"Lҫ|tXkR0k2D ;ATJ{MVi rVHRybk kN& i -4̒99 X]dHb`AE!d@ #͹;|$9XD#[iPzNJV(ڛ04e7 ]*x#=~b9/&}Ud@L|nvXƃژVAWxx,*IΞ{q5W~V*׈8G -hfjjfwpmhl&c r<&(;NZA[lN`Nj E0bWX2XM@ ӣ3ϱ睗sުz|.LGgnj\[{;NrC:4CosSDsև8;ký1knGyΚԀZ*C(vfcz3\=iT*rO|3MvMcLpp)_P%( z&$A 1qפYB)H'Cb&Z" zb/5ǜIPB 'hcІ{$YsngGI D SQpAF"*τF&=%+Қքy(]0c0)h|II,D\@ꅤ^X &+u&z 0&R,M/KԕW]'ʒ2i/ rin ftDD @u>S|8ii?J`M^bk%޾E0pks($ @8/p)∖yƃ/k;4*[?uK))뇈ii>X` 7hW31Nn^j[9_l =^j_N1^PrHku* Y.6"ިsh>5 -8iȍ8=Lo"CzQcat&OSO$wڞT9eWWsWzx[e* vG}}dQVjTQN a{ʙ6lӽrmWѨ2l?]e/{I>-Sn*(H͊ѵΨjҨ]NaxF :ń7yvg/}χ±Ka"uSS?7zވC6:Y))gI19R6$ż3 |K(F$ iPԚtd"%K]rjS4Z^L-Y )"1$:jch2p:R$5BiCoN/fq_X%}!Lsq`5p)STxaޢYqJL1z#_Z2E rkI_͔i_| ꁠYg Vsl( }y|qlb@Ft'$HC!V"uHPtŀpY.dD`HsB8a%qT%Q p)Ml4HN5>y~pP#lu6q\cVoUy3fnΓ_?LF,_1+y dRlAϧ^ E-{kʧTl7}G 0|18IziWQz@RwhǓg[{$IEq^iOfz=UvlS{0JK%Wdq+ voeFV2y (0Hq>w7Hq.F$]mo#7+H&w6H0~&3ؖW'7bKcV[-[1~)>ŗb!.qEʍZZ/sr3|S8δC ~k#'Kkxm q#Xd:@8ԢC7OŹ^!mƏQ/G|.Z^8 'i~uK6}s{YMpR-,}x q AYm~k4+L_=` 5lJeO$gK:7C#|fUD9 q!w93 E QL$`\.ϓ_FҧrN!10 Faˮ F  K5El8`r:ߗSovH{v4aLXȱȦ_F1h/?m(Q)~2Mw5`S _Lhsrqty dTr3{9>s?51?QסG]uzuG-Vd\(LyX΋-q9ynɝ!eB&/=Kߒ?ގO[ei%Hq}62/947%%(R0. 33hq-xYJP.~bzDmɌJr:frtҰh 퉐15`IFEABLZQQ˩!4p$R~/>Nt<hָY&Krs[ {gۑnGs0ڸ1Ο~A9gI+5=1ܗ$|J큍&K5WBE1RdJn fs' cyn2ʝY5jA;BoAўY] |*Aώ-q3"mSHϞ{ p(C/18*,ıHAZv4p$)UgQ#pU.f*Дg#rc5BKlAg=^f%8a]%-`uIcoug.^[fi;upqbhc5SUIM=vrQ5&b] $',^/ls8 g ŘXћfijͪϏ8OCT/e+&UHyRpIPAV$ӜT [`ss=q@V fbaŸ1 Ua]h˸v8 kfjI, "RodN2%Avil'a2ESS"xgzL)Zg #ڌKm| dzYALRT ' =`<-`"fhL){P;Bpp xBLwg3L莇}Az;;7-?8u,I=wu if@פ [B_jբz`QP^O/-G| s[1dȍn+}1s TDG fW,6 fQoc0{~u"F2Ji+V?Uo| sr>Gsr/]b$nQA}1>7c9SYJN&װ5/XUX1fMwiJpuWgCpk N/l h3tHjt_а˷YȲ- Rl]]o=^cZ2kWe HE-1${ debkpd}h@2e^ʾJ^\Dr[G @3ׯb;N1;*4YWssJe,Sx%qԡfmbc(p,g3r1dE,"eM!<@M=h|%mB lK~UjTIUPOPYL]O.X˅?SPXȘKáe)vcӛPhOn( Ft~PBnϊKg_ _Gw3ov KǼ4bJqb*7z#o-va - eM1b n͖$x>#'roeymz?$c(N%Mc#:ةK6l=#fmTdC oCm &\{`F#  r,`8cU&9xB'[4v$CG0z25/ݰx8}S&>ݽ ܂@9?`lTOL.' qV IBu5rR_Ro_w?ACޕ?D=w &pňC1C쥜wg\86j>U8vHK9 7rʭ1z9nq!b*/k D]_{#mȉk54A*6._GM@=P Ê"Ww%d7~~yqmIƀI[5*q;u6$]y8b~hķ)cF ;tx,+UZ/F3O 4tň=ԟʒyj!CQCPM&v  &TڪšK{Wa }ikGVD*H~is 3]NQ}jiӥJ~صNQxy[-6f;DǤ-.!u{VK6f+y{ Pp ܯT*Ir~xÏ; -gيބ|.d+䕂63# +ղ'[13[A8jG?eͧv(@JE&kFxKE'"XPɑGĂ** zc5ʨT?H 8./*4(d.E>QxTQ@' 5s$)-T`Щ7*0DvmV6;$nп'TOxx t ߾ˏK3*$!o.BPi&24bHlJ0> :臋 Ⱦ+/F5xV)+jXJt{7J ~:P_-©Nr0{\|*h3gǀy-JP'c!>'E sjj*._Fj+(Np&IYn >VcwprICϋ ,dY 6gAhܚLPuD9 3i)Dt?Χ |=ޅ--FbPe4@4BPX]eFREaHNd2s֢ٮڰlaEH'`%H("гjքj}yORa5 =O&a5nڇM֝5 {w60/O'x˟狭tuQ\F)X9l̙˨@Nt+gnc])z?E'[9O̖wU}ҎNTJh5<6. Q [v*Q=$G)fWO 5MTMMM։=Nd&G)f{ݯ:>Ͳ7c{JXs{hPBiB(|uѺM qy7p&H2c9+mX{;E-`;++c3E1c3Uh)hHԲ3pJD/0,?8Bf%c=9}%}q7kCGu]_}TR9uI#QVzᴳp)".DGǻ[3P(K98X-4 t4BαԡL@} 7C<(J.pR/'$}F)޺FP&Wֈˏ~C}q;]jJKM) I[zxg <|!8$zT=06%FlB7HIP PeW8P$1-aPHn񼚭$Z#֙i";}aFP&R⊍"6¿z/!b xU?8#vMP焚^}r3~^tr##;n#CH 9FƫZ'НKɎq.A1G߱{ቁ?[`Cٵ{O Ȏ7ΎW> QvtvJ)Ur2neu'/"l*?ϣ"C= Bqث>s8f~)c" ,q.ṢDΉMF3c-PDs&~YbЙ~)h4 Ḻ">N_?⒥o 3;qS%m3nA.rde\嫷E>\xAV\d5N#8F5)(cА Ӎ(7=d%ӝaj,>wNϩq>@ j[V"qKJꨋQR׋ ~UJjxeW.]y+oOB(0kfZ< #ڳnvs2Qml5>kd{|H*\3oޕ.@#鱖,J0ÜVQ$Th!th`˥ЎY-uVRˉ9V3ش9k+ws]R1/Vr)8["5&+r|a-2i-2([N4d(mښ9F3&`"ikL>u=HFwmMJݛ{秵̦)paE>ր=7f=OߔDjgƸBpE1GZg"$˵U9B #gViFmo`+h=-7.;T].JĴfl!Vsܺz2T:&NDJW6'\p=߽z=⻿mbrwATz__vǕ,|={/*,k_|s-wW4 ?5џ7wf:/'ooo?]]t\y[Tf,nc,xkd Jq՟(I@I~1O8<-rZjX,2RNyT.N lviލ!dWm$I['U2v m/ 6Қ·p>w҅+2SG![c>Hl;lT S#ҳ""ͫagZ8"GUag0e6Pdۯz EҜ!IUUcg)K~+ǐȑv6]V5<^q .^o8zܪ-Ėv,+utA6sQmm%u![gr2*DQc1p8;8k/<$'.56ˉ\p;ǔB h i 9/R̉:%HQ>Z!eYNOR~-@1?][O3-•۷ $lۊU%ܪ{q8Zl%9. *z#< )fF|{ ޮ[3aL/ON5R]4=V뺒ޞ g}. -4`,6 W;3{D$+9I1@ vtXԴ7K<6}ZU!I&dGr6"$gP,.:wE)߀PS#AqY!Y'V2*" juҬ*JFz§7@щi2* :&b*`FCpd({4DQfAOW hI5 2%6^ȰAb.ܢ>Mo`li S 5⧓VW~$?_۾銝˫syul:ws P7iKpՓ(je6KչD&%Y(fg9 E!gDO=.ڜj%)/M6n3Yf~z͔w߃j'uosY[7( 1ϵǍ(X$D&E5'Gɳ.p\)Ԭ%F#$ZլJwrf.?dY%뙅N E&!?,P ]Wg[Y3jpRW'SKkzU+{N_N?i\|JxxcF?},Go q,: r9d3n)/|>ۛN5sg({͔ʛ`SsyZwBfg|ru/ǭYޛ[+Kqy橣ehqBB^)ݨ; PIGqNvF/kHtgt15)m6_BB^)1;9Zysj{[_Y `ь~׌;Ν`fƔrM7eҳ(巪ӫ+-p {n2IRFNLvBE"Iy8Ko^,A|]a߰T%{+F=7N %ypi3|}Vx3b]!iKVFE_nthm44i Sȑ8s!hç`=/[Tc磗jMcAb] `=o ,  4替Z*ъfTD <scdgz0, 1N[6w'%bB/E2i V3z,kQb4E/FI''% < ֳ6VF ko T1ރ̶ cHب!٨B61@F$ٔ`PdGN4ս*}B* eFu63\՗峍m,6qB9º0؝9B[e;6Kojy4Rzܙ1l=YX5=U}& 9tmrH&0ZR,YCK.x -N)

x4YrZk" ՂNoGG[\^\P^`w\\x{fYeSG-V/fEN5:w#^޵y}Co>@W5p!]eڄlF-H'|ӠG.XNZ!U/|=ҡXi{>k_e7q5m.W;Љ,fmD)M RyBTQLz"R%+ v:8fCיTC)~`Ȉr V\gFMϵZFGbV[@< jP9'֦txvQT)W?Z,L6#NkkANƮ?;ZAR>`Y'$Uc л'D;"8㟄W8_JaR Vy SZB(%[J!TNzrZi;0r;]mc\q C2+aL,Bt(:Q!JI"c|R AZp&d1Pb`ю  rN>D$uBFNWz%Y-J"h!XR^wڠ-?SKiS]ss ɟ=G3ݸvܾSPW8y ]ή?$\Rܬf?{j盛zj盛݁Cw^XzN"+#"RDR+ yEO~vƊWW'?Ll%{O9W'oga%Hn>Ot<{{tv3{͇x7NSo,PYq|;!0vľTĊ#F 6K,cG}:8:/QIk1@聝6U`S3"d܂5.`hM.17$XIkjq\rSO{Je9g"# FY)$QJdtAk%FSI飇 F&8-θfA:~и:,$hQ FR% b BO(ޫ08`*G)B*C P4serT}d/- krod4+G>\lY~#EFknk#$2ޔcKc1ВЍ~aV/mR](ZJKQ&aH˘\I!C] Luc#+9 %t"HD dKf#*6 yi,J~XaEq8ϲ"-:8.)qWq Bt&Ѐj_|DNC`q,:8.);S576~,1D4C<Y"h]RgBuGܔ%+̐j)Kԍسʸ8+Unro/W`qozl~}Iz~UʯE I95w8 p}dUf^vz'\lӴiy>ٟGō~_P`l~_ ۘpX}udV˨Kh0^qU.Q-MR hU\Gspjx4Nއ WfE78GX:DS{ԯsRQDS,V~Z%!yG(<@Ǭ?Q N9Ҿ9l>ɬ6-92K;x= q$ ׳,H1pVf v!(n} з ZlP57(y~M{ƞ|(DDaV1tN֠P Y1l@9;սYmC͑8W7Js| pcOuhl׈]"wXzu;[-8aiw 9'ib  IJJ5Fj&Ĉ`pij6 ~smE>;Kp CV5ڗ7Fvz3}eS[@Iv8+*β]l' `q>F ViBʔajAF (&fUDK8n$R~gUF2Ti YF{D5˰iR!;$D?rl>!S„kR}\ <r$3!wG r=!=B~nA!&yi" h0D4hFv&G[ $rtVqFNgF{p8SݾI{-ƎkxW)$/B1S(>w yN)8N'51rڊ4kYҬ (E< Hr5!np`8zP1z:P^tT\TC*|1 N)N4DZPL#YFKID!JE(QiP艤|Jhm/[RcRɵp}C7L:]ʣq 7hM/w/ 4;z(nze 5< q\ i;Czw;dhZRpPFt{F:Mgj#t! }hrzhl=tlm蝭TA[ ՃDGyvRPџrsYPwhIPQۻqgB#R3ŕs/TXPNh B"hoGCEw${vBb:eoWqi _ }|ay.m9iۇudGkY|5'zh&} JX N(*Q(B2T4L |: Ē* #P&BA׃ٵ̵7h ˣۯg0=!nupL1Ôb,Q\&: cF$IBI8՚r[\"K>x٪u[ Ւ Y-joء"yC|{AQošO݇˿7`HLԦ=8nUYh#l5|x3LSPP1 /N8>'-$D|MiV`J)UU:]wŗ7^(~ r2_wБ[ J.޶ EݬE8*Z$jc^5*h{mmEu!T2ZD<,,UR(' ;iUs tB9=u)Nw)׌/ljcE=~Oˋ]\$ك[.F_dfLxMꍁIDm1"3sƋSesHe9-v3J ]nK[=LOFID\80aQB D',8Du)!֕cSDh L,XӠAޘ7\*9NxWj9 q%{MpBhkOhc?2Ta릆Zjí* i!xN2c uUo=t%J-z&U?xi4Ϋ5C+B]: (7E^8) n̫_/P<M$nϫbʰ踮6C,R1SG0)V$Rl Wy|aofIhw&ھpdqk&݅șdOrlxx6&sS&Y3&Y/).ť :EP8d &b#dkFE( EP,2Ԑ($2JؓPqG!%IC9RaM} hS+ς~o? _ЮHY,/恈bm8ӑ n\̷.N#I~Q21K8&H"#^>$,+B}Kuv?d0ڍ|2hL1.pu-¥f{|y%(}fk4nw)tHRa&0&N! 4aX?b-e@*?hI])?|($Qlu0pt S_2LEt) yP?8}Xmy7"db'콲M5ETa6ݜG;*qe8JЬENYSz6S!3P+x-vݣBkVkXkKx{HxgZPrEV|6s٘( Ǟ#AW)ϵΈСS6۪L9̺Bj~`KVܔ"ؒ vSa%c`3AgC~s8^Uz3Mѱ r0шo{S*ZWRMβǯ9];™ i- )b ^POA5[^J UNÁ L 9wvK:5;\8dj̱%u]=EٻFn$W[,s9p{Ydp/ocٓd#%٢^"nJVdV)XEVRq81RT}XDNʩjmg۴fì3K.gtev3N|)@`irn-X! /{/<ٯH)5)b+v#wv*$ VX)"wF} n+_8`sy׸_x7s4!41mVm-\[G-3 nMt ;;aհ?$({S'_;WP Ueۯ\xM/Dըk6H۸ry!3RNߡط*p˥kp M ͇| #I}=o?V~oheHb Xj5#X6Z2ˢ qD8f x: 6LRǩ+eV("[?/.eBlr%[/q(ZX|Ckeoks6j6c:ͬc7U]7Eߢf,|u҉?N"%III[:i `dME~!w16r1bƖ ANG~ʉã409$a v.6&`öy-LpJ>9 ]a7ಥt]lzH,N/V=BR6e-(ҙ2Q:C-US% 9 sЦAgDoHnG *T䥅g<>/ k}Tq~{Ni۟CcdBЁNA]cb\i טwh! xyd;/?xn?QaƼb|a_u&|:'eR?N5x0dgż'cLfsp6 gen&?N+;trfY4P!e^ MUQJ-&*C*ATP:MFhp` n7ZmHA^1MgO3/gN?]?W#1^4TZo'4T@80!q<(%Zd%Ս ל(8)8 s*kƜBMA`~ȋMA6Eb2}]|0Y$0/47zsMhO,`g K[[/ƜZچAjA"l`I]U!4X,XFvQW` R/vDKB Y@.g v1҆d 3˚ׄHQJ>ɫ 8#Y 86BKf|ֵ" Bz؞(1piVjU*X>]7LӘIL5)\b u!rf)^g C9;d[Υ @rF<'( ú4wUXQW\ifH<;2"u8R]`L( Ό S .`9w衻vX?a+1r(l̫J9lTik@;dT(ݾ}4v]F۴p92ʡvee( Pƅe4R6&+״gƟ,TӈF ܣl^_`L@C&`nj*5Ν bM7o睆XXKj^&ƒEn|9D|a}(`D7?7 ^ <#]j|J/&ैd*= OC p7W/ZL-ZLz]{.F!8 @-2N^=|&aazR-w;2 |xJS%w0 0&`b7OFӿ4~׉=cQe|nrO]r?~9vs~T)ewEl?,n^̤o ﶛu4ȸӏ!>/upv!&uC4bJ(~8(>@29 yH-H>a]0w {}{щ {ڢ&"zY+Υs.# Ldvv z-tnId1jHυ0-g">D0TSXP;Mhdd#!Xi}Gs;XΚϭֵ&8;X/ބJZqQ@LϙҾ1=NS'pFҚ$pmT(@ޠA BjMޒQ =z7au0U*C7n=8W5$^( ($Q rf_AZ| AtX'`4ga@e U:kBBÁ*zuTiW_~.a$=ϭxui8$v9B`JMrnk+!J&9+*e7 n0L%Shy&"TNXqAhMB@YVuICkcQe]R^+AŦ MW U}]}{ݚ\ bgf&\s5rw<) 363)cD%RJm7Ւڪa $@EX"+{_kN2rSDZH ~]`%h#Mr*Aۿ|"۱j'os#Ah~`_G`yŢ|ouQ>uq[ѯG@RcKwnҺ/Pz8;AvcеVA}iiz^CX="DӁa (rP#B ؒ⫴.+_rcEpB̯\ň_P7qߒِߺT.=@C%!Iih=ZnƕkmCfIͰ(}`_}T=%l'ynx b8xGAbMBm|{%7xOp!)uX)3L$wRwNo_j@{JyKu $Mx!582;1`.gy,Ol(/w圲ў.~wzZ˻jA`g쌁%Xk̷lw/>cɞpao^ٯ>O:u2CMJܛ&'[wήLQ<3yCӘc Fn]1:]bƋNh}[1-yCӘr2]R7P"!m`֭{ u1B2IEM5޴nڑzD!m=L fݺyZ#jzF+ˣ' xf^Igq̚Z J/?KJm2]nREhղ ߫QXﲩ$g,%="v :y$\-*`ztGQ~oQcF0Y[,;)?݊B2]r>[J**ȯɨGtu )BXRQ97wE|ͥqFAS'af+=8o*wʖҽ;w}N"g) %EⵧʘhۭXu0zX.Q?>;%{qeFBGafd.*3#)rYȬ9MjmB|_O[+O⿌NΚ9%g3rfLWSq?T뻟>ߞm{نycy?y H}g{ f$Ka~15OkF*eS 0pMN3Zd\Ԝ<'EUooa2[1|q.p+EȀ=,^9"-wFDz/HpMp<駇{Z,U2{4!B~KS4糘:΢cB ޓ{5 VE"5{E@e'/oqKgƾ#)d$6J,ڼc51)#_ʙͪy V!ĥG$cgSHV)o(J(WrYRUθHHcSx0 J-Ws3L)VD(J; @Ed9kOlXW׮~ŝ l) +'X)j. US7-T.†L xY&w F ׫MV=O?n‰:v?~zkp~8_!Siy-E^VZʦ Ati02>|yvo_@aY*T3N5eDUsGuo[opÎa+IPoGG0HԣG"L#"P>r*W'd]ܩ:IOw8jV%L ʉ=YNo)20/_p"|UE򈝪BXS2d?Ӫ_-'b`mM~UՔ9YUVƛd,e4%l5ƣK5}(T.Nd~^:=>mm`DGG#gUzxzdsC#-ifz[9R}sMR *d|WD'j+}܀fR8@0>ݺRG3jB1Ds]\+u&N[ql>aV1N W epktV($gRR1Bpz;yݜbF+Hxůӛ2/ݴ۹\dY?#(I3LJo.1!3﯂WA]䫠.UP嫠#fȈģ63^F*3Kxr!!K"TDS PF<NQx߁WAۉhw(B0l(s旱4r0'"GVV}8!F4n@z#t>H[է gٴa> $BGmFP ~x{^kq"6r& b&ieDy }gp"Ҡ"eNM tS?[JjU |I䘠[ ]epDR 3Z! CyPhC bL0 I1!#[;ATuV5f۶o׹R ቤBtɠ!*Pw+aӆ9.)rȇjڻ+C;畑)FY9C+} 2c1& `wpH~0^0ZՂ^JntJZpF`r ixaW+4;\I@]*2$ BGHpL{VY%jIQ]Nſ'hlv?:XfԨ>ܞ-7VPTe?x[;owPL\6nSՠiA -dHs%^D:|1|7=Edٓ]dOvQd:|)IޠgsX&Sjx<:)&X!J]vVum'/?&7mӃ0}wl7Pc )خO6h8r+C cQ3 A'A6@P亡md\֫%).`b °^J6'aa8V1knOQc!y@~/bR0$r#\[2Z1 _mmr A)N SWmMuX )0(dc30[ъό =":2RnQPnU"DxƗ"9B."Єqx@Y+ѭXahZ:e;)?|Fs7$D*'L;L!0JnsKEczbPĊ)Bjd8G[Yf:K3S"xa1CL{DADeV`Gg$ƴӻxd1Ds\Ӵ>~N3n&>K>W Isw{ONb 79;=m]zI Azl5C9g I lϑG -0Ve`!DZ*Fspt9bŰ c3@qTr:x6d1Vv&tATkzr.xӵѮn>nP֠vĊB|&j{w8+AKԉ@"AHަ2`h[TP5%&hdURR%y0+AAb\@JDz L)a/SFamB#G04fe&VXi&ѝ4DQ흖)UBh1Σ"D{N#-hJ "L`GZXiToثbRUUA0RV#cHl L׷HpEDvfEÄ&evުA%8L0/oKV}QJ;t " RL;=X%b $LVe3Z51ʊifǫ/biFJ4"k(aVJ)azDV.aO"*K$fRI[2R&A*KfaS L1)Q0 3h0 yFR*tHh 43oՠ ș;+G=ä) Ci Fr">A#s(Js& "E}9CmLH08ln1a/3#,kӼa?>cؼ Azq*x9P\rjӲL|zFfh ~ThuތZ E, Dy0ehgfOlNkr=BeK[d>XJD.oLR #<.|B0F=h"P2aCts>.)'iĐn~{?>\緳SJFb=@PiKwc/On>'1_vr6c{5 ǙD?pnJMMV"~]^T?v(1><1=p_޷\2OBAb 7ll='o xZŐ *8^X_xn> zxH֫4nN|Tek/w8]L6g mXY3O[YF7?v9h=G^?WwC(oxt|9..薅}e##9/xtQN)mgZ+k40yb2h u@9打P.|i{;7ng7(lKGkc﫛`uoВZQ1*PSԣ~G7W+z_?cJued[#˔ۻұiAzOS0W94d̴6@C:^?.ȣʼyˇ 3?NdǗXcKr`lymEJPƛNޤ|y28 =x|RCxv@2^lB5 qsWfj }t>\x<sbOF(wv^ ch/KMC:0G%94"lN7̱2t7wBRZn8rkԽٮw;#֡&[z.T&w.> uɊS?1Ö6l~uO_7c3m _CEߓ{j!W}A*O߶sW@_O܅C]Lҙ~.oxXV0+Ae7wHÓquVO40CA,hIц{ |:X9GtKٵ^L3&K /x?6\F{5cQ6䖏t C:%1Lv2X$ӻEV*rtGbZnYlîe*353߈w/%0bv"CAc=P& \% k 7$)i>$_jϋFcTQ@RKA8%R(z^vijp+g\`HSr/a(9 xsmgvp3,΁A>Ӱ/ON?؝.>*Q\t9_(#$뤓DC.NuM4As5պX kA[_  ߏ$t*[qx`ZY*hBd:) Ce5DnA1 :RLӤ8!4%I8 08D(3`QRnvRq kA*@jc4;g`@ C zx&Ea喠TY@%¯ :_ߐ!>hNSJƕN7CwR=1}~*Au0fZB|85zxAʽ_;hВ5 {V@Z% !1P[2tcŠBoi3&( tmDݴCy6y +Uߵ}pir5ܶ:Q+u>\ZFZۃf mB[%bwè z';Sv'.ty~~  u j >^~$9|]Xe3z+mfۇdAFG}Hpg-]Hvϳ1EetDVczΙd=0OqMk<۫cJpN:A[ լH:OӐ;p~]R_qa3~Ӽ6vvo<xaࢷwX&Ml~6\ (VjB*@p6mbqw5UT׺CdnK앎؆iD:"pz-E\d>k8/>_[g;JKdξku):@=_֔VM(U&mLUd;ɱf-L}B(eH(JD}s␢VRB:{.h<7Bu&r], iTءaM(K姼0*rw{]G/wpH:jw}iP\e* 6o߀†3:HϪj_/@ ڄ5,0WDנ%ըLUXGɲTjYR[";tiDYf$džaD[ָ0ex`:/^} Tt-874GDy~c/[EO~0eDWU$|H"!SXeCSZc_n\.yylǃǶZE-h+ G&VβC |Pj a| a2£ɴv5VutmΜFNЫ{S8k=]W˲Spdz*if*Qr͖0s5[UNEN;В3.@ʑB0hU`6"l0%Z#Gr̜eiJ8|.hf[־+q|>rЍ?7.zLɳK\[1~ T˽)?EvY^9׽8e,r0_JA2 E#tyirI0!T5vR\l (X>7`+I *waR#c !TH+qkq)eրHb.[^Fa͊SH5esLk9݀[!OqwN@]34>]bB[`m)84 @;|曨K:ӻ:\zy}a>~Z? Cu 8)x99zݦ7 Z?~rד;O8=l -K;I9b&{"Z`P\}i'E;MdGt_d`ƔoqSN9|lx1?ł9]VdDհګ'BSu)ɇs$ҩ&UP7wo@!}{ K 6$Rl! ~% pv٣6JZxHA䩝pہuEgrJ'eP,;}&8pԤNh0*flQoY%@u{ZV}2*hP;@6-_ŧj}}ن7g#?fMDyyfDn4X $ S0_b2-fdurej3O`ous[ {O7YcԠh~\Z^oTMv~hISdm0/-MHV=<3oݧ}-\a|)~HW.I2&žB!hT bD'MNaJ4#[hLQ$-_}<7P-$^W^c|ٸElv _bn 9B3Θn/L;6B1dW/ n< ۵ @ GE2YaO4(ԼO'B] 5Uu_A9H boVkvҀ2Flz?L $POZ"@%m,aMY`s^OlI"ynɚV2LXi;)DRzfYPᡊu0%0,cen\J T Aeʒ>cmXlS1ϬߦPwEýE(̷+ lO#e<\.hO1 y]QyǹCؑLC`"6HΈ}U?Sڕꚃ3~&w%_=+ Ҳ{F#/Hc!б`AI"|n]p"&(ONu_-=Rzr0Ϧ7Ϟow4_ܼ{sK>W_8@TݻEx1D V)vii o;ElSSsQ aK5+Qcǜ%!4b:<Zp_D^lu /yCޣ62w[u'?O|іuv,|⓻gz<_V7>zI5aO'{x;~|3sAkj8RkLi %89:KƔ2D¸0s 0 %j뙛߲g~U285'a}WVIʊae`:*E0E>ߗ R߽1f0 20Z""K~c\–^bI0yx ?E$.wXG߮o>|:5_hT́-,YE'h^yʛWl l|a#1* e"7@ÿ9Ll!XaOr$JLXaFTY3o ^h V[F/z;Z]`g-%K]{kl_~f|, (SEpֳj>/܌E3rycX~T ̍@L[9fD^ UR.n<;oG 3m44^QVǦ̊{] <86YhGBdϱf3)2‘ȘSd$ " O}v ykvx!#y!/R3=Y8` 0:)AL;չa8ݥ&/$N )'G&)"^8#[UBǕpWK͑nF+I^Mt m梄ڛ1N^`昔 Rœ)K Ja,2 `B\dR*ɴ0Њ Ig4,H[J0SgXm骶ׇO5/W [+qq\ه9Zð O(̈^|780!?!ko+||g?#V߹oW-bsB{eVa|{tO3uDK xUgqCLHhr*Mj 㰮/G^[ sTpe㞇2T≼HyQAes[ps?fPRpY-~Zl,jnacLe,g֛w(E!2+,)OsoD#XoJUKo|f/xrŻm;{!LFue2ֲe^I84&֚GK@ެ-'t}6#A6nGIH:CY>RHCݡ6M׈=VwLɯQ@׼cȐˠCG;P:W8^qD kQ)Fh8;ӆk8F]2LpgJvBv( [?mXPQ0X )pQtfpqpj4$˝*KE$HKSY'pB`5xo6xO] 4h8ջ]6D)Vr)KKP"%v!THhťH%(-/yv{!Dth!$C{PAS[Tdh!.,__JYk/_k ,lnQ%'DT8OЋ,*)T>A7[o t>{nu X+apSaw1'`s#s8`!7@Q\cy^ e$81fE.O(邛#CP͂SI v};9.y*h -HW.I2uFDT bD'M)Yʂ]kj6$+ePsc1 qS-ǬxO?sύ\eDJu}fևGDS*5l IYEN]ġNQ8X10 '&ivq̱P#r =">#OA£}2dFf/:_a`v'f1PKZn4Lcx c((4qf{}إHwfpiard~IOg@-pP{C՝x[pϟbɝ,N?L"&5L &I uq̠J^9$0C1qZfү(} A9Ï ;e*SI\aThugs8κyidׄNjWFt=x#*)\1vd4ƹ%+1fY;Ľ"k~ȍT/'H u&D+F n9Ri{!",m.I/Vbʷz%XnB-R-1CvQ-.ȬM,R%|y6ggj{ȑ_1en! _d1|f;`͇K3d-yfC,n-vZkgcŧEȪeHkeh$O$&)Ԉ|ј8Z-!Vjae+*X)i 9)RBZo킘LJ`@9oӁ80u LU% qƂU! 򌰢sTE<t:vHF,PNx!6<5BLUcaR.L!V@+kAk%"L(*t %& Cƭ$[TU6piIPiRmTjzTj }DIc3攐NTjkaj|)``)k%9:dyC!+c&(;Iq(w)cM&PzLZP)`G.`f$B3*Z|ȄKq̠N(q<.EEPv/u^}D14$qz@&Y`(ٺxo-3.E9kz}"E{5:! Dݽ[9I@B<\ÙS`8{]۫En:eXz1sok/4zXiHKF;\» C.: YLy1*uSX^snNaShuȮ##cֺ-  fUW @,!NW 8$&ދV>WnuB@6嫓a^Rdcu *HZ`E&"5Jh4i":D ŖD|rUx,{FI`OC6IyfQ05TG, R]Ie v&Q.B|haƎQ6Q0)g0!H7tpq]ՆrUn$WT{9 *U_ňV ?/b#4znoK-ky)9mt= M:׷}X37"B5yd-u씬dQBщc-/䲖`!Dٔȴ)1uK-JsƂiTk?+6_%AwBˉLh<*z ͧ\/?}kEZI8iAb/kҲrЖ%3;/uݰVj}qzW&zB+X(Q(0`(U$ t4BƳ}["Zs^ŗ׻Y>-_9jNTkTJMD8=صqW-)}JT翼n_O Zyh]k@hLqSC8|zLρp٠󔫮Dܚ~ɉG vTQ~2AyfJ(QfiRHhWq7W; 8@.!}Ͳ$a.L'ú.XR6e;M֓:ΑJۚg 2euv:^*e[$%b &Z~X_K'U~~6o.b3W "oް0©dafUB\`qϪ_( vD+vx|q}Ea`b*qB{/XzڻU~hqU9Jn tDlG1 ۛz,lW;YP=u] Ȁ(vi^)# ';.alB Dwg1R`=vO6Q]~67zKl_ײJA]VSGl)PmW ]~>?8ևͻY,>|+˓ߠ SQEj1PZUTWDvht6žbd9F]r_nյ~ ZxPª2{10N$ L5X0[79W"!Nq%o C@^&Gk0Kg:06I 3C=gy?;}4 Z(FF~h3`̡zg&0fVtAi$O{?y)F`D`v-2wV1>)=j2V JC08*(3(9"X en` L%NmAUH5QUbP_Bu829?xNgf*Ψ1j/ _{F\!RdaWu28 Aq+I15:گ Q9y rl`>o[ҊV\QwSYYsx< `LbdGd%0y|eyj8hԋkcVe3P,ڵX)e%KVR*HsO8Bj`b^U1FDl% %$IAf\s1OȇAx*W;Y~86?\y["zXnntc5X-`V%+Wm9bY8g/?߯A.2A.䲾ټ!L;P"JEݣ@8 R8@&u^[ ^~kTcUb׼Kz}hCHZaSacz<|qǛRDzݯW>} /_6OkD z/jwnb^[2$utFrD Bbۚ Qyh/V|wgʂ|ZgU+ 5;b T50'l^(@ ;Uϕn9LD: _T7Y3U0 zUꔲfVtrNPơ?JB؁Bf!yѴ|ƚYnV{}?{~Jt*= b Rn;CƔ!|8q{w3wY'NZ'odE6YYP4r7m ŦឥC8.diA!7alpnV`̾׽O_>aT۷boxbt)|в(L{!(~7 J eWK/<ހ͡! jLˮ1~zt6WE"Ԙ{h}{O?DyDr)oӪGG` c+u91 5UaA{TG 3&@ӚxGvQSLj f 8槺W~R\'T  VdibUX+*VFlZk*ODbTXDqZ|@j$J6ӿV: V-W 9A*("օ` Dˡ)ѕ-hzԩ?ۡ\LunܷIy4d ^S?~tQs?VJXMC#lV#.o^vܘͻUBUwRTz|?1:}9A&䫋kXy3k|b|W^5OjSHwl&+u}IqL#}Ԝh,O֩C?snK)Lix[S})!C)ѸXjDJ.`5Qn7i6VvNe: V=UO?w&?WBF 8]Rg fs+q_\@Ϋ$QOSGdgPl1&_iל,LqFF<61 5O[eMiN0!g: Uſ^Yx O^ MLAO 8&EƃqJ+Gr!J[4і c$89F}wvVi-AG=%D R *ѩɼ\ҩ ݹE{gnEy.mKT}_̊2)){*؃xz乇f@{c-x6>qI4iF\B=\5Iע崻ǧb܃RjF@MNBѱzg\S٪K4|RxzۡvFq3+>pff:( @#g-H֎mrFE5_rYUqno튓/IM 8[YU5HDIH(vO7 ϛ*8w5 b z>.IsP(&L0契ʼC]e2nLJU~Z_I 9uIZQCt9:0'[0ӊPrYi`$PKb"q 4T#-$BX1d(CKɋBKG:HŔ(R,003@L"G8MaxHjGX"bh/gN.tH$EadcT>-fy M6f"bǣڐ\0kŘӌ! )ʐ Э*RIdb9\4G JJmJ hh"ufSw>PjIrKbLb>\ˊTď;PL ӃdomL>|8 ⊭?@>ME{}&e8K&TZ R J2G-[Eap"5DQ)n|zrOQR1rC2'H gQk^¬DnX`-($Q0T8DJ5 D_@>qo_ftb܊,Xm;4l\Β&ٜ J0Av84ӹ=j0@rLuMaE7P4sO9kh$¨mpspS2iC& 0;᣷!+f?vU>5"rjڇ O-U̙=sE'o4EVl {aCh!PXNRR?W`b]Ky2aq6Y& ޭv0]-g=xig/NIl4^>r(f6~"sk@λf{Si:)%>e{ }Aj6%D!ș,Ce +\)ڋ+[..Ve^p"T JO*&y/6.1FTRLH=enSrOf 1eKaKH)K#\ SDfTqML (Sy_[~n(9r@5xr7Btw\KTCck(:YfiIH^,t`^LMo+B4!iD X)ǚƆٔ¢C0aRJy)M9OgsKSjy^+Syk7$o%V q,^\W^e|M]jmoRn|&Ux<[~&y}TҢXfפ%Z݅|/ǐ+_[l?&՗QqugP1} ֝o|GGxA%+x2@=bVO6~P_Owj1itJ%TZhqPv i1GY2-^/ӭ 7nyyRzkV.bD[= 4ɢŹO22S|F<n\Y0 -.Z u>"5LaBo`Sm彩jE$e1qmS4Q4"N]*RTxX}RK;N %ppK ,Vʼt=*dJ\Z+mBxNF2#a-LBh,1nhj*97%oV ꄵ@N>YjwiK3 Zq#B}JR5(ɲi%>Q8:3D|~3& z}Y5V 68ЦW VK9 rBߑK:5L"Le MbN%{Hl{3Zv n:"|2ͭ']W|"! y&dSNލ2[(>:팓?C~ӻ5a!oDlCV> Ũ:Pb~ifEXiGVdNz*:v[Z0JL&e47d&:w_>Ø9{v+P`ΫPO_UcǽMW8ץ$v/؆_QU:cP ޻szr & ~"[&yن: eKZ"n5^1 <`>^CYGZ+\j\Gfu=Tc;A B UÇj^^MM BꎈwPM':WMq?vǸΥͫϡ-EǛE/s-V^ *-d-3ޓq{"t< , ~Ө&T=D Fep H)hQ8S2S{p\OMO}ԚRmZR'K0%ÞԲʡ08iЩ*--& /hM4Ħ3Iߞwѻbc:ӀNy)mQ_քq ҝ۾{ ^ćbc:ηq:-ߴ݂;ݚ7nA6%qލPB11wn#\w,=[Mt\|::ͧfyuo5)x{onG0TTcE$K !ˇ4%]khV~.y %kRbh$@cʒb$(&1%3EReQLx kaŰ BKG$IOdLBBxpj4fT8 sDXы@5. }O71_sF.ec쮔pxzr~=(}|y8bJȇ^>)=Țx?G?-y " 儿G. |,`G{gBǂ ;V < `WJE|{|=U2H79GF87$X) pZkNȆzJrz!CF,2#ݤC?{ƭ K/N(~qqNcW>l\̄rg+}$%/#bR*qd _zzysݫf$bW㑽Z5}C<(Vmgu^Ku`ۧg>ES>ESc+BPCr+Ű)_0"29`sV2a$XlMfm(?㇭\މ,Jsx -CQT]dF>CvI724T_x9_WTu#l#+430a:P.Fɡv ,P+ʆV,X/>v) ݹ4OACƯ'eu"կ_&7[IʩhUEf}~hhJ'鄣K%7a/Rʾ|[K{Mu~Ľ~x=9EƾE^/qZ/. Tr O9-G%. Xs{tqBcL.IrE0cqa SB8;+н:F9`A d爡2CKW@ܻ—A[W?ݚCPiǝg7-.es\ZV(Pۧdhak"eƸ'"` nƅ433@VJ_8;OL4‰! MV-N+e d.$*|9SqD9 %=10Ce˾}Xtf&j݋ԐStMGUʹiR{eKFPP\x-`•1NL<Yv0AJeݱZJ"o,a"Њ;LeOEY^H∆1-P9_N#~%6T%*  JPrl=J2)d'9ޟEu$$CtE뽧qY8Z|2>nRTw}W\@#`5S}U0;/y I Wkd@~e旱6<+3Rvb>7bjtlgᝌIcz1as4]!p/70)wL`}Y>+}%ird[*1.7ɺT.,k?@ rAp͚*ֹNyO4G'RCZi<&Yo}!}P,ȻVf kp˞LfEa >#3WH]bMAۜ1=8)8¤ ߤ=ߛgulPC%b.D0~YW+z ~|f<9A8A1N^ʱ.wiS?ǵcOUKhU@7t}` l`-uYeWi^*r#pUcEk,;QKC⁏_;W~1Q.F}sK3ao;E;WJ5!2'PںED4ֽ[3Z&4ضC* "v]V{/[r8RQL5NQ;\8>.xr7_Q F1ٽ!g}/h+/kOPfi'!SE @e'\ /( X MVIeKŏ7݀ ʘw>]# |dW\{EJZCF DH|WGYr^MaI:eq7{lze |%, _~(x W.4YL!L#,BASL)F2YaRj"ȆG-8q#7d u0J>n SB!&0YHF,*SmˉP= $y ZW,e28 \ob(PX)kPky?P&ڲEa^2&[u6E u1I3!-5ם|)3V3m:!o-8*䌖s,~%854Z8FM@ƐI\C.(CtVI ?ǧ˫p<*aozb'7Z<PTxc?Ѓw;=S)) x<L'kR9(Ӕ(6ܮ G?xlt3ywU!UlR‚uPS=flZ4҅q:8`֙ײ6(g{VƢ=1Hg:Hc^ދ0c"Ľoa=Н~2:0%f >.>ar,gZPV'DӥcA׳;~#.:«TQ bVxRcJk۩}Nf&bšH.N} L;LO[$mo\mTƺʀj'T?823(FԤZKj\؍&9}BQXSNFtl2i<[ ;{i:-͟JV&uͭ as-٧{^.`[ؙog9MT̗yC4kL {kݔ}n 6XSޠά[>?Tukb!5h☢ ާͮJGaUM֔LbLݠd3~T{cO]dMp *_F@nÞ3/oGuv"JI%PJa9)>N5q3Ő,p}zuܟҲvjS) fw)1? 5M?NW#-5Zµ|! X!llm F9l0x8Ţ 6we4 Y&/#_1tG5"ݱkPA 쵇Yp!CriC;aN;Cy1kD:$`K~Nw6{h$7_l:G%귘=#?Y7J6󫠠AW? hzt4d2t?iUq&+Ta7hk7cwre˵eέ~1zVGBhQ4ik7KuM#ȶ_ [#~uTC.XG6= : àL +OHk|lph !m{RJcN>WF4Rqp_ӽP5::Lu/91v.PJY x@˳Ed~q0unzys 2 oꋙ^GjUmJzqj;JoqZHK2OI_f39E=(d!3Z9Bs-lxL0 Ҫ"}bMxϞrĵ,lt2 6z vwE|uS]r55.B[ 08vvU(8옕|-(Í_tF Ct܉nxk(.yYo w͌wͤ2$a捜"I0:ϬmguLة…R E5iy) DP9!l<ǣElqwbZޘ@XS߹[d壀d(c.Bk~)¨aJaSD +-T ݶԌj"\ RXrۨC$y:tJi&$!/|?!7LXތdiĆDcd )׶g "/ FyR"yZ駌a ޵5mc鿢ո_\O'5+̼LʅmmԗHj;TZ$RE9|GN_Wl8IcCuKŵWiYkC6FF]wE{`2!uFL A14~B jޱZBt>|}聑@!oIjtP?hH e7`hh%umC_brq6I~5J!_ҺZ3oPJDB:cL9t*;V : $s%3HSlLFNr E2ƥ@q攆!@LHo8*sO+ T'$,L;u^ H3BL)+qF |9:9.Ӿ?+9c@:Vi1Hrh'+G$k0 l6V?툺uH9nfPAD!9 ) Wq!F3bKrL-XlYnO ou D+J xz%cD 2AvWj^iƼJ}G3B꼉kadD#O>Dی)^'8Tgš;M9YͺhubɇcJ@zWk?\4FY;^1mQ_o9;F+DʕaAb{"C lDF9D@9JWk] )^Ui~g4IG5fD=`1eNW$qEu$[V\D= j''9ݕ&Y>ݻEjY=4hˇ]*a IA]}{<nV} Vh SƨR,530MJkQHIuA[w#M"H|<كMKhJ p ߄53(ωE7O_dD2h *z THB[ Tz/0/Ps {>|̱)4b%2SL2,)9c9:](QsfJ`VDžw[FcJD:h:\"a( dKn>Wl8>x+e@ϪKS&'9@%Θ3wV(*E'4Vw Uu[V >Frʭ"}]i#tj'd/ VBqȧ?Bf!-<)~Y᎜0xSQ1@tjNV3~ 1gCf]L$G =k=RY<ğDiVq5=1x~}pGCOx<G-&lh`_3f2Wc hiO|kW7c8kY|?/D:H흋rEr(_c)ꐕ89mHcAPOs0ٓ]czʸ3C1-Sw^)-fD%H )WМ3gSc9CEPdNM 9 E$VbLLG[Xy7&d7+ڔ5 [S"!J'yB;JX#QZr`4bQH Ut-6$`*!dXIL/+Cai@%*'k+8!p (@*DrnUO>?x^Mf0\Bp 3lvJKcjm,J8V41` ŷI-e=7đr`irFONM~khIKʧ`g6`$cJs$m@߅tzOy"kDi#*\n哕 ?tdh{yy[-ɧi6 t_ _yKo݆'B8a[`\zӉ_a^&GN'H}Qƺ6s.:l#C`m 0eb):kh,ld 0dcpEX,">BU䠑bv6A#O(Ƙzgul6fa͢'2KoHjCmnl􄶌D[Դ7>5¬Hy"Ym$+ee~m5ҷDd $ _Q M2ӏ .Ar2DqY)f"l !F HtD Li=Yql8r٘0LI NVB eutz%W8BIetW3eȿfRGt! fV<#g8kNpL5*: Aå9\Z J`8f,{ ՝T$r6N{wgTQ+,& #LL2Y ss)C TNf̣+89@ (m(e_$:!EpW OF#InUacz+vy* ƒ3'LJpy;\jI7|@%rFcchJfλlNajK?f\jKFfsl| )NMrPKJOHEN=^9tM} I u$ LUOw} cd{tˠgk)LCI+-v_Wʼ4cݫFҟj/|}[%u7\8YIM ńV/]У1_0hUo|%A~[gYQjy+5xH<|/ӱJ^lUpI9Nuo:%A:kp/y&]7}iErEk.>k#QkI]!q){s-ß\}Z/z&\)W4V2Q$ǤguIT;G&1@y$RY!*Z(hCxl/_̆a}mZ~1Vq[ Y2lW7]EXp n~p'Rc* 8MjP#:#F$^b \Cj5LPvN +DjB~UHHunXq 9-޵@{yzpTx|P…rS;[|-YnmZk,]i U<܇{wawt瓏޺ *))K%kV(=Q]MR?~+SRޞ²H|i5^~yԩ9߸"ەi%$,C#VmYvB˱˓& /E%ASL}t=,^\U"i'ꅼx IC9Xx_[J*4 #2D8_Jٸ/ O&|8 Fw}\5Xus6S;zx\-Lzʣ;ufHFw,%z^[|1)Ca ̒}4??f~Ggm_7E>-;QPNzn.g3Vƿ_E,%Da%یe|nS8̲_Er[ bNw41 [HYn]?m1qUI>H*ŋI8r+,ﺝe[1Fdu޺nwLv[BFvrQ vaP]-a^**ٖ.M m%G颚}ҔA֌5$" ϠFYԲwa@u uֱVjqA sMfeҔg wr4#EII|cԧ|8x 8~ԊTWkAG$@*w?bq/g1wl⯑c_H8z61J)=="MFj% O5T٠̏42xu}kAjh `\HN%vg[C o^a'":0ůiܯ1xsp*rס.(]ê%c&K\c$k8Ĩk\2+t%&TCFJ-ANd'4~ 2{Uh3g@r¡sƠQXz)ʍJK'0H+BqDVh 82.s>@2HkBY ;J?Z@`@~XޠgJb( R='˭t3f]ђ C$q)k@-УbAx@H %J L4_U8J~> b) / rӉq/3<}SWC0~ՀM35Se9(\8%ERx%R &wLi'{޼/@x5zlhFm_jNCuPb۽;_~Vra#zTq<,GcC֭lfӉ]6 (ջwN :Oc%4݉zm̮k^3+R jHzSi &FįYn0;jezH2P%h =ofT| 7͊;2HSNֹOGbރ~hN3PXtnjۀ́9 Q ,w8 ӇM`~@0g ciaWOrp 8/-ܑ  ߢo^~~Sz`^4ڏ'. eeՄ ?/v~_)tgZ{ܸ_K#~1)^umMb[JJu"HI3}6p5s{2DN3)mFR]Ca㇢gd:8x S@%t,mD."Ju[ `MG;; M\.gPd(2Dqen\j (1L[{fgWLmO ©4xS~#!u$/8¨ 5~A*ϕ_l >d$_|3sϓfJj 8<7#$ߪm^n ӢQ,qg Hjwr awʡqЧǣĖt3j9a2JxˢPɕn)1PC-09J}1_11蚞.%tzg ;Km)kCQ^kծkP:Cy%jcKi!}z-U?BμRpP"I- o>V Ɲ /;,eJ?3y#)vI ΍V^!{g3WstJ$lpRFío8{z1mG|1`T` ٢EgI$KJm!Öx x! 9 ^v 18]u j>Jz ʺB7*ڃdNPEš+";B0>*MW `?vi辳;* )u,?p6i($4W=֔6C-xWSS)BSEK ƍC!,eVͤƌ;K0O3߮+d(;PE6,,(݌aV߾_وۇ>7o:D0Md6M|ڸ%`s`_1ϯ)V#,>Z|ʟ*?}0n׳OS~BG NlӚp16@^9ᗫYV0%9II;=PA6t"uʇ3{@)vHQ8Lj9,o,G{^mml|׍.c(X!.@}fQA4Bk*HH$e6C5 due#D`'^~n?/@;0đ)F6q#g$D nY=߱jp{USUPPz$>K7.Rr(#&4Sq:b/xE+M#&Q>4#fF:!<1b9 `9PA%_`f^ɿ֟eL9Q"zfqnX0"NNs9CP:)HWE/{  PoqCJFXcq3Sl?Y Qw_6Xs %bVp6}/2 /ciY)ڸqVY$]w.ȟERGȿ-|ָފFlj/͋=w@|1oy_6GKE۵/aFT8'Ĉc 6<-ђܡeLȮe;AqV#ܥeW $Q2:/\^MטI(b%8i *J`G1:Q'cbk q%PnʊCf)f d$ ‰TRH1pb!pb-`gyBkN ぐ ꅧ7 \B%n/=EVJvki"dW+)|[ku36*J K"b* 1ٮƸ"xyQZE˗/^H 2! !Ȕ[:Nb80LC0(F!VWۮO{ r@ `ZʺETsƹFX"SZ79kK!Lv/0џgS".[S7n **_$}ã-xČU\(HZ'{ß;1L2!}D`NH*RgN!bER"GE_?yz3sT|xeZ)p;rduHvgN]/o3ʋ[zG  K/j5> ux(0YˉbMRj1i' G@! ,{5:ۼ^j 4<Z!VV 17BgULk#"rbPCkm#fbc֓ybjl~>5]ι۹S[~ %g:s5][=M 6Mȷ2 jX?OAwV=Bp`W @ h3J%V!=^ WENjc[R˘^t9&o7AXcĐK}y Vo+(j\TP<_W`S^~ؔZ6K!{ls$=u]ӗ'\Jvl!Gzg!lA|a&{XNڙCtVT[*!,/-VmpX!)x=nB&;l|ux%ިϻ۾^`v=#['?Ĭ7: 7ş4i7 j5[Fqai5um-IBÐ`$pZ*Ś[|H\V \hi I~Zm*^isltמ.r5 RN V\BOcqij`BACbk f`@l9TP X$Pa θ:T@Hͅ1QA]7Q)zFȝ$jG2YWF.?2L R \n Q6$5/M:Ϻx" sk."g'naFh ʝ$k].cgZ83TjZ-%7km:ǬrCN8ŭӡ8b%$s`ܗw6Lv@N!83v\ /'!_8Ä5?r ʽ%_]_>a\1C^Zr_]x ƋnFae]Z&*"g1; {E/ qG 2eף۰!owalPPۭVbuG^m*N]%v VĘc).i;yB [!cxj$ ʓ% 0%u?cSMzTxU#OlW9cC!-ApH4[x͝zp{x5iav)0h虪g hrJ0xVPO_{Vd^v~pY{ncoBqLcT(&*6q=SŮ$HI 6V>Bc"tC~ԭCau~4yiDj3v*>7{l-cPXBFURĉ4cr!bj` ךqi$vLw UALWm8rSDj]x(pdWdV7;g+Ll-v壚^]6ɯήx5rCPrV[|C|zyo;h}Ɨzi鯜ތF:NR2]KQannۦll)lG} /3/r+QadaQHO9RWfs-; s syk D[v"_& yw7_Q+7]ҫe#"}/^买Fzue?EDv!v;xa!7 ܒu+w/ֳDn5{pm1Dv#m|2Zbcܘ&2h/7mc9´,ę?~+q9Xh'4 *"ZH(#̤" 8™D!"VY+K^_#`2¸=oL~ gz^8GG*gwܚ69.y72#?g l6< dM'A۲;;,]m:/u(JiW3LۦJʎlLbQJJsذ4't2ڦBZb @#CJJ0mSyYzba2 /i!Ȏ1weq$zj1^+cmiA#(KR#٤dUW5HXőqdFF%}Y7FTyĬpM @!hJ[.(A?x" . ~da畀1ߑ"*a"[qB]@ )|nr@d,,9V)9ք)\eZVL,lXdcwG +66Jf n)ɼ6@N3^?\s-Wg0X/OIpR敲{K{;BG0 اJQ6<{q9ztCoU㱡x8nnVO8 4jMސ֓fwj5}qc 0|ˬW|<7/fa8iC:)dkfe|znl_̀MxeHS jj-5s7_k=xtqyW%)D) DY1i5~ąAaզµ@nZM$ӣ{K;15WW[ITJ67Br.+5%#:k6RFa$<`a Ox JhKCƺXH&;#k*$ۜB,TD>gfBgvl TnkbDQcIErMICصpj >1wUw`W!e$9o  m2tmwWIЇuzC.%D P8U] g]2n7ԍ&C*o}7((l%$͹R&ʾw6ISN>FF^]HW?ɲa<]ҮF22&YG+4&V]'iC4>g/} {n\ƭ"fuǣvs6NfW]KoCr`5*/IBHfX@hڊD?Q/ph}A~3uՆO1@{sdo5]raY{VÊ1;ȵ_ˇOLf-Oawջ>uop7eL J z*R!y[lņ_jVC U7H3(.'Aɸmh$jZ!Kh޻?jHfOe_48|ST?R fIHsI#5MU%,![E|cZbPqQV5Eӳ[ۈsʮ`~7ⲧ`in@O}v$(Gy_0h\eB9,Ľ Xh7T{x'>u1-:ߺ]8..]8m (ë1&~hn*wT˧\Xd*4 /~~xr]<6F=$9=ॶJyrY>橥Kn 㮃j3ed,0SvYO%emܲEԃy|jHV^;/כ6vgS780z䑄x]_1ӣ6v/N xpyp=9cefSCm{EQzH/W$9FS;K_QTA43;{Nޘ.HOv[S#'rpqkd\ ORÍ,j/_`3l{9q\0f\FpOp$wv侥乏=߮o(履7*egp#g"ROԝ9<ڟi ="*|r_Bب?+moN_s(49ܓ|î6c[F—#_21gpf(@ ~ǎjIl6%)ZPtΎj3cHeNG2[{]y]djVHZ++; 6PjoA(Zol .Ӫvwxwbpst$93UfM M!;DPC*cB-Tyjk"%UY8*ȎUostej)@PH+FN643V#)kjQ X-c9T0u@gv\js=ʂi!eZA!gҰ h #Q_z!8m`]\ݸٍUq_xUOԟ1oi 5Ij?ďm!λ'_ࢿ"nS/+ȯbot &!m"NQ ZO~J/fﭗq56GHa׻k_R?˛5~9ǹ6-57uxZ_goneLC~bRKt> -Wj}?T8Y81fl_CQ׸ o83aAss 5lS'akP{[W?_C{軃:Xӷ믻zPC9UY@!bƬiDfGo_|[<{u.WWc_ 5cI5hLe0ԈTQ? wL9h +`yTnTkfďyċ0H&un CL! &k$0cްIh 鑦}.ݽ4#{L+ehe9{ƊFfuc^1ipM'I=KM̂gnZ?[9jZ,ܔE?:>g7y9;sw]-rE~z^n?Z:o7?KRߧ=i~0/}}"wx@??,:F˯|M#,8oz!!YRlg%v8{ Լ9xZ? lY 6 S"[1A }lȡur}xOi)M!^ .LE:da p2*jɖG-##yŞ6[ }M<>Qhl=FtzܵrQ-}ۨasnڳ~Z:Bneo?;9ز+ ܕ.> ծ4~,o](6хdh赤fđT1o=R)w9+V2[ko# r$#qxqg9 -xԳ sw,b!)[c∗S^ ּ- ߭ [8cE:D5p ™oNXM3/]5$k0pJb~;#7,APbacp\Eŭy[JX oA? KYZ1"u8$+M(KʨKQZH)$!*4* 7j;cߞ\%j]cOPվ&.BÖv|U֯CSpIp&>.M3S͕d*&BU j@܎br7Mc %WIɫQLTU ZIoƑ*T TydHd~o|,}3Ě /Wh^/J3Q~NЈ7lKsONpBp 4hP{IpǼa n1lqD7wPk|DM:7RK 9Z*m2ao,\j+ лCx깾dWս=o e>ضZNL: {c!k>D?ZCfnW }A+\v8jjIfxa#MhtC WSٟO>ߨ{_[Ǐ|+^zǷ<ٻ7n$小VdeXbo/FSM:kp?xG#in=<2zXbG.&gבϣT?=ACF@6B9m߮FOO$3yʩWS#7oznӦÓ,!̔.vѼ eל\i%aV˼ͺ ht.H|xG"fɇtf( ѶS!q OXZd&qOI}x α~rF orgǏlumnGqzs5)SxW_@.Jnq9/Nxw0]@&u*Cfz{'+'^$6=KuI9Z)ga.QXm=љylJ%-o>H[6:]cC_}?JEW/h# E R}1l~Ǡw$[D=ztOLZ"btf]e< 'f";֏IÑz:j19G:~=gԀ0.+'ll{?<[穮5jTGK/ ehw, A銄=m*#E]Nr MwFqdp&@VƩ2ֻ9BjIܿ -3_ c%V(A31'Ya+/A9^OiYv,a j<xwk 6!yb"Pf0֏Z=zgZY&9#Vʁ %H*U 3$>LHҠ_.^J#qI>aAhn3#%7X9V*ED]V*3zq#.sϺCU\BWrbцw|!qJ6NE#d~xN^*C<`)n^RZљ/S/Rc5BFcR+,ǸIb(`ޤ &`X<øޕJ`LoQLMYj?BYR *> ePK$2Z(>΢J'_kc@!IIzLU1R4TcehS@4%؊5eAHYo*ER`By."<)R?p#1؊VS?"x#V1)9t̸E%s㮹*z;|걳Xh X,%FPQ'!^ȃX,ghR&svaCj6h-rva,mll͍Rfv˶>lݏ6Aȭ1#lG j)z9 1duf.&.߬׳ {ZV yf~ xdz*/{)Lk㔧p70%#zd0JG6O'ͫTYN]=|Iu#cNK5y]g" gW&WG> ბ2;:|GHZk~DMz>'%Ytn_2ew\t{idzǣwf|rM傞-ʭ>Fέz\\cG+Gj; A{`RFd>pf@ -EPA TYɤZ h{dVcŅ(]Vߎ(ymV+CVLuEmɝg$v#T  A"F0S2\Q=WEgSn²(5^Nzzݛhm~A$@g8w&^4T&C\D Dt41kz \i;NpȾ〮N0%zJ "fdR*(_EP xGZZ("B"g䅕6FZ/! 1o2G`wS'Cew*nM-~ZӰq LⳖ4dͳvNW!5zŝwGmIL$L+Tt-5*-qh<&O5k'LIǍb6ӛ}N|zLBrgoy4NS# ~xz47X"Ozn¬\pRuotvVS->y#$䅋hjZtne1X4hcӧ{nF3[EL=zu3LuH,V퀻%n9ɵn=H ѝeʠwG^y[Gqh'> ݱcz>PG F W콄bW@L:{sSsKX}s{p[*~췁rkE >HMi{v˺mNaEs.CۀZ{K{Q!{&*j+QtANBBwə9ܲ@zhP}'Xm,Y!f c-L2/cHe?$Xgah5yݼ{eH*+X4+}9#l|f1(f!2K=u㲧x#fQH6oWQQg)=a)5L,O(HvoWQķ-y׵H6oWQx#r Hƥ-SgI`VR9<}))Y5XVP+zбH-+.2Ҋ @4D%X%VrwԖY ;QI[(eɍ1ܕPe Q*GXPfǝ V -Vc@4\IWcJHWQ u5ZVPLPCGWgЧ֡}`@;R>)5D./rjRB@Ho H-Vv4 Wv7Hw62bZmHkM%63me.-.|emB]6]P VTI0_.Yrn"<$fr6 ).{D#5bzC*VJпXZٻ6joo~ѮTJ%uS\TKq?4Lj"!đQ th΄ 7{H׌t#h lJX49O<հٺx?u_u*'ޕ=;[)b[޻~:]:^vhu+\WJlTR xF3@;r ݹ͐He΁g RRLʜ`F/$29hU QĄ\oY}ㆣͿ]o^-ls^c96QҢ+gTȇ&/L5ZZZ46Z ^J-U#Zq7'>S?t7:3j3e tJ:KY) x|?O~v.?;߆8ggizCC< ą}=PuTF|6kvOiz䮮NOwcHt+!mW,^6 V_-B> K QjHHsE8~e/%̍SI#~&˄/§4 Q8ыl:@wm]J7]M4܃^%9۬z4n|u,rr*Y7Ӯʣrݣ?sMo -%V==򹿥iR5'g+N牵.!&ndQKjKHwYn/ o UWi=[=UxWyc@`AT!@OHasӒRV+N)8&uk(J <Ń*-6! 4f9-)ԟ~\^~uC8򶆁lF hJpn = B_Éj l+ڵB˄9[& ?= 3"!w#bJg -0K D-Yƹњ]HZRjW)՜qIe<3-IIǒy,HmR2'|/-Hdۘlr6_>OnQ~v =ty(^Lp;[>w;l9ϧ%a&]3>Q>\Mot|wF3`PqGo piљSȔPι~4FhcF{=Xe`Z\2Պ>Mï?z#ޥbB,}QRً~H5ŜPK:t$@ly@}OnH:5%ֲ̠\erl,JH>_3²ZKް*s= F׹`G8J-zTnnWN'rF/EQFnLL"Zg& uRR.9)TkeO#ʜ2"㙄R9Q%i)G?#^ I1'Kyi.Q2]L{cȂL,( Hf] ؎q[gfTCwm$Rj~hST_9Ԍl-UNKiҧ~HQpmR]iK5q% n$M^);v?>MAٕXha=xfS\0b{ɥwuϝ2Kb4Nqe${dUL!ָ'+:>{p+jWb~pSwvqr[B>08F-2ޟ1֟W.E6ݰ&qS鸇0e<,{Bz x `)^OT$@~F6mkgʃ-mlWGd ϶h@t|M쎷PڵA@= AױI9{P!AJ/8H2JKtH\%cidٗLZp7FUFJjǻ1"R!&_Ug(oS>G͘I`U 4Md< ]Ens"Tڋ_Xcݟ݊ĒMଶ*Ob9'rF#&eF9H^8QaωN(,`_WtA%jѣHS,rpJ3e*"N2 2XZr#^?RB4G(͵jE$` :P2Dx1.2 PY֖V,u@!svJZ`I*#$CںcV7'c׋ vEx!bFGeNŇ rCc21ES,QgGmmE˺$ XwI҅~a&I,nMnB @- <fҷI\J=ܕbYI)$r KT$iLK4L4dS{fD6˳m'QZm(]@CiऄdS2䞐&4,H`UJ!RJ<§?Lٵ^Woǿ&w6y0Jux  {.Npsm4m-5*R| B>_" XJ4N+=#V5Wf$ }nxM# -z @cvxr4`X5V9]EI+Fgi5/ dd%rĬBDmdNgZC2^s~0s]{<ar\ޡ͜g|K] [[kFU2FRIѪ=S8?YՕtDqL3҄=D^C4ļ"}1A|gMɿ=&.9^3$~?]0~Hr˵?~dz|<sgd/=/1ͳw`c.GNrL]|J#/_APTqbޯTOo]ZIeX6BP-x$+\|n(.}{3WNٺ1̎S䮒FϩxI#Þv!N,m, q⽕x}%elzkOV+ӵ1߁^ dKVr=c} jڝ`]ɀ1)aBYZ3frfTx;z'G7g"Gt7'p1)^~obXWz,c8-X-vq{71߿;k=;ޔE}D>tw>D-9qGX<"9\|4!Owώhឭž'iڂr>Ґ\E{锴jtnغ'u֭ jT;X-bX۽E7m5TZ!4+W^:a?\.M;n}eP:}źmcjiֿcuBCr][P32_Zw5#,Ja$В缙ݦNZ5KC|2|]{g}DN>Dth[Wwb0B3JL_Ŷ3<}S1[ ךAR u$HU_vsYea;i+x7[0h6WaF{J$t.'77I|C%|ACawmƄRC",wPQKꋁ3\vP;tO)?;!4;WQ/>?b(Gn/6 !+%vڶnD-[ UKǹoh/6 vU34m[7 -|*SRHfu>fsoT%C,@5hOB*&4)a%fcql@bţ'PfZ$nc4P3x8~u\9wGwG#A}s. &( (/؟[#R$|N%C\|h%&Qjo , Fo0:9RBr1̄r.q9^"X$ 4d 7{9'HYec&խM}0/r]v:oU's TPGR Aܧ)~;v&?Oi|_U'l#\c6N ͆C.RPA%`v牚]d6xGulO'9~JXC(! mH}3V/gkX"P"RҨó<#^H_R:bjs)ah**J՗sѲ[ŋJCRPCIX'!He/J*q;4_A)4;bgQB+;zz^eSRʆB-R񔽥ک[s7[hcP.ϲ.שEju|u03;[[AM2ǯpyijD^WWuf3wGD,{ [ưxTEt9Jƪ6E@t1vh2wi>)۠+u'ǫHMѭy6qTmP $ѣOp q'KAƄ&]-Fc.fcVeݬLe1ճiGDz&F Csʢb_A#[;xhRĪPt1g! i`*/^LGaB!ojGG9}ao]̾r/*_CnB&%7nf,c㗯wAOʇg;$$C] 33hRU4m;e鎲XMfh5QTUvsSZxe35~7_BbÊxLb=z!i:Uy(:XUw"ä*^,_0zNX -g^,](ɲo0$ 2&-C!a6Glr'@8TI@X )f놙34b/ORIxP289 {ՁFvz9A~_ ֞jxl_ncQX]|㏵yTWG]r䇻4m]9mۏƙ讋aYayu4#HC'F Z 1ߥ+8op:["Y!!WBcD-G8RZ<>c'ÈqI:b\^91vaBFy#]#c}Va=`q>Ιj݉aRY qG+LCIY%drHԲOUN1%5n(-0;+A;y-l|sbRM8;  A8cCdzo"IG@ H G9(Ҽ*:M: +6=ߦVi$e޻jQ)F ;r*}Hx2` 8+mtӲwMt` /6̚ `K\ j~V@U. 81 HT tuƿ؎ /ŜWЖ)Bt89wJƄpT$J$PaS83%40Ci8'PR@Doos溡TBc{g>1{))0^u-:gr۰Z5Ճ6WK[MSVV*myV/nuW5W ,aĨD}f:5m.Օ\zɔr7 s]BldHDNoGJ+3n*_}qRY%>m5/ws{;X*˃dUzϫaf;aXI;%)z^Ļ^|"PYo(iyB9I #P.lneΝvXP|\eC ZS/3^e{]$ƙȴH@dF)޵6ȗ-`GRaqwAƦN7JrbLOTJIQDUt((jI>~X5rb1%g."9\ti^:2{暘 "PhF;ؽEs#h0?̓!H$5Ղn/{uL 0SQ=ݽ$:H 8I> <ӅJYˌ ,<7eRyRL2*,UdvH͋`5'(nNQ=ߡT۳3=D41f%K`nY3)Tg)\e#P%K)IQpO0P/ dgiﶋkVc^悕Y&12Q֙fFSa@AdERZTy6JkE '{g1D}gCE/.ɿi ͂6T[W״+#Pk耬c>tČ$۴$ص,9)Դj=RmG$w<3 }хko4 G.DzZg|fP# mhmd z}&Mo  ؽ)Ϸf&Ko֏+"_r]&Ar;yX*9l~qjK糓g'N:ټr~Yu2,2뮹 -;;G;s{Mw/˗KUUmxHH$*K5P .Gax'SZ AwM1%P±c@Cj>c@öoByĨoč:Q3OSmCcBnXZ#6y#H!˷z9uf\ ýpHK{p]h ;DҺ/ՈoN2LjDB{B:wT+!(w"yn9F`)a;#;-hW>|߷nGaby:cXdLU؉c֭ *SG {m d@Չ#ĺ :B؉֭ *\ZސףްEaPC4$ khΥ3h%CX@loz=3*ixp~\}pU0/a\>[UpYxRMXE.Yx病o8i)Gܵ!VTȋO-53B"tغ3^ӕ Eu9C,_&xA-~o]沰󑱀=OHH .S"b4¾HG${C5|$A<6"ި1% UO+olszh_&߹~e|xrUX"[ZΌSw|3}+JOV_7? 'vڿ\&uRVet8-E+!M%C ߻u zy{ Z bzv熌RAצgḨY2M-:[/L22&Y/>Ŕk)XW۠1snvx}HمrI`g{LMI'ŝ+ȲnVI`8 גϼyd%Xa /{" ;e@#q^vɖE?%v&EQ~  FUbl}H͎P\=8`cgpmZ\^GL2(/ReyfGkrK"}϶̏Lfd?jjl9|3FQٗdZJa*7pNZCY%W%pqkL)͍Hr܆_ *y\ՔQ&Ȉ F&<) Le0LdOI*r"cGN6YV́tɮO N-Ri]hctdNv%ʽMʽi 3Hԁ%R.('E[*z˲1JIhOe.5-n9a= %aFd1STw6.ݠ[KQi) |/FR"rf-=?-57 N=mliTNNqh)?-&vK\Kil|> -觥Ĉ.ZzZ=uKxFOTˎZ~aQh)z.ZzZsU3`Z]I_zZʕrUiSTw&f.0筥TR0zZz.kM-=G-৥fdv{%/lytk;8qvgVSkGzۦN0Euaj`7zv1%Ġd7|I *uXrtIA FR &..U"笥VUb ZzFzWU$BK&~s:s-TowFOwjMP<|y.D$"$r Z 5-do@*` 6JVK֎Ab+֜ɝ~qkc%ϧrim=sk 3E2,]Lr];zqv!P/+.{ldѸ}"zE5%>/ |j/&<+y\앧 8Wm_x{QAl;.k:ƃA2Ap/c ҹZ :HWZ$@vV>`b*3<ӒPJd&R3pS2Ys+$):S&/DsJF1zqcʂ',J4,YQ8cij,H2y"e22V+d1,0gIEU#C9mj42m}I2W h4 rXJYvJgYb&Bl Cj=1tu5E8Fw7hLiHB#,PJrLW׭h#r!2"R0TG!=d/f&!As~M cvy`7b%׸ZI iP^0k8\ >YO[`7r`}jfZiw.0a{RJ;8x S Al-`\MݽduBwלOOWDaQ7/Er巾`5jeg@5ǾeI5-\^/Ρkaw>wqRzry w^DS4ԺRef9:d=ƭ9ϫ rCVXU q/?Nn6dܝ T &1[.,%-`j?>ne}v+[Y+kljU, In#Q-Sn\ JE" cR]r^"+QQ*Ddn6?UagJ|m;J+'PykO|.6مj㗢6hz.w3LZ)I9u@wYn2\ OB#()RhiiCP9zUqnz޾:b[w 3d~WȕEm?# ~w)P g L$KeAI#K%(P#+hLs,IQ1RHdd&{R6R_Z_zxZU9UL~]ZM0,T.VߦU:˧7-PsB]AucGZр?@_|x| ߬`y9cd-X;P Xy 8N'z8Ч_fq\ {s~w)& si:jrsn slϜ^@&xV̪F4TaD~`&ܒ0^=p&ģڄ  !=@RѢoN6F62IjL klNUGzw/w h+T2-j}>Ǻ{Tea:4v\lH'Ѻ[|M&![_9Ⱥf&} ?/V?kw%:̓nvf%qFM nbU}˺\Xm'So7`5Bk9yΔZmncM딃4䃫hN!F Y7#`by:cX;+SI7_ZiА1:F1Z7΄Չ#ĺ :ͺxZ64䃫hs >~bJ*9tW[ ??^O0*Km(;~qLdgvz? 5GGe8mMar|R۝>nkc>]]K׷z3[H^z3RW3@ O( IrInB-$K)HQ&ILeZ*eLctiuZ2S.)B:àZڸ9 0֚|jAX*\s׉dVW0iIrkhYeI9 )h&lG5zb aum+zV++CEZbaq:ZV t(?G-C'3PvoH5@| Z~g*g0c"Ǘ>^W^frPo5ȏu\4!k=1οox=[1@";:\NC`hnR+"Oʇh""D$Qa8zEKOg1pw_MaCćfX@5ؑ9D %S7A,4d=])a,ԢaiKSR$o{)ڜjTȕ&q rm8 ˝1SM5ڱ.[U&/'3,vm*9ޒ64䃫hNIFߧ8"yU9FPYp}1Egյ!\EctJ#ǹg4֒n6)&֡>l`\3/.mκ-z3P(:Su>o8Xg nsU8GDOS=ߣ^/QaWrLR|Zɓj)E hKJ%= hi^7p32qyνw iYL?]uTps}m5dF[/z]*\|N/xWd/XX*ErUemobWrNe>]=3`۲b,VOko%sՇh@O瞫D_SCkb[9| t;b:/A5ľi}II%G)QT3E8D<9 ߙG,2S$!Y 7(Q^Fɖ5.ٖsb.] @iK3Qng:sP;=V|7Li2%M?5p+ߴIQJd4[Sj+rҭ-^ֹDq\r4U +H=*5c (\!e!0OVXǘ( B(BmJuiPT |ss'*s_.j> j8T25ξu7yxy9J:}E^xX)k=ڿV,҂gY: d!~^o5]r9pEi A}nV !@lǍ y!A"ǚ%mB =Β6!)kiH!YҦ $vPWsLP 9,EVc/?" daM1ݟ-T &,#.o}Jj KjNgm{ ߆8CL1Ksr U0s( >E6C=،"144{a9*X.?yLEY8itK8C :|էl(*Hfs3֟"t[`$:UoTJ=M{sEq͛yl/bZI ZLw{$MWެ2…}QvvYY1CT Dd*&&̫ *Y>Db+@4̶@)dzU;Pg0"9hhgOaFGp/FǞ^=QJ:I֡~Uz+h{adXb6 77Hv.]nyJgs)=]KiBPH ĒkM%*!($IXs(1h=hhJ -Ƌ*q,9֨\668Ba Ʊ\"`䄉X0N!𐪄LQ>9WUc pV]ac}ĮY/EX,*^ >0cpؗ2gccH'sؗK1q$z5*%oC%r>~ m7XW+*zWToWz=%h[ \G| Zc߭Kj#@@8@D"(bBtwjgo;Νk B8 8RQl):~pߐ:uEnTQ+TA:/lAHf~X]`hI>F9vJZ}|h/\i+c.%pD$L)r&fIb)PLĔF1Ds# ]uE 6^Ce'b|(9;?>;4Rf#eL(JVa"$D1j~b30,&*$P=!w!s%/AHSV1ADtLHP*I u̱đܖ.;ɐQhɱlwڲ:"`fsev sN%/`2bC@$$$$H%\E&RiA cH Ŕ %:VoZuX_QW'ʴ\ӓښf[7k1SdBBT3by2v udFuZ5RV=[7qtIw-A<);oj\Bިdq#;G'1lhZW1wt -t9vN˝=Y&@}OJd\(m;R  ODd猾0(C bԿ 9°3xyRv&#]០1=QITP}Wܒ~I}4*̌\T&W~6ֻM?Mw({OiGjhDu(Vwck(cCA,QY, -846Vu(zGGO0-^ 1,so}owx+qKg -מ7WuE _i{[[RjlNmq5F'zR#!"Q$tBZҼ̷/oYķ[ŢT?"LN0Y-2{c d|"/xgP>^\᎗S Džy4U;wY;xW ٯ*66?>؇عhddGA?U]Yk C Œ6uwm^ =Dfs  \Tu`)ýY7(wGֆ*SMz8[7suW.:혻(ͺ#:Z64W:%%y 'f\&Q\O /H*2TQIGʱe{tdVSZ8!vpdo'kf_t]毾eNn`%vk XwAjF[Fy]!GJ\({c{ 7>A)\ԽB&'W@ hcܜ<ʩثM>Uc! mnCgֽZݍBg JlSk[MaۼGppDnwzr_gzFA%o=NzB-ƦS)azBDۇNCӽ_SQp $vK)%ū黍e3_${u=:/}!lybPMdq仩)dn:C"b __k TĚLhcHJXIns>U0SZz3Q*o܎87rd{ Z8$IR\0ER{* 8S 0~V| VCI%jd:._f| Bftpk-3[5;>Vcƛ6^qijm7ngZ=X>@d_W+Ո=pTwу}8mHsD;5gJ'NP .$yhxIiB٦Sc%JY:e[%2VD\G 3IR8N4 L㘂f{o{M@y0gp:nfyv%[qZveiikpAKZKtR$murT(qL-^I-=o-%+>k'?5gAKQK)vRbNBKv\AKVKA-ۍ-h!ƦӉVx2M*^LrA@&#GxxA X6SJ]1N=7[lf@mݽ\ fY݇j5 ɧHC$5:J"ԉC aXslƅbpk>Bކ  iDvF8WDSRtch!{6 Wpbiھ'05[?F6`yi:U埾49#gE6P;pų$3[5֟/i.T1%3mo>; غ͛Üۃ,@hej1paxB&*d?;C\:1 $Kh xHGBI+$ͺ%M=``3:l }8wCn?rvʍ''w#@J$9H1ɖ0v`[Ih `A>h(d%aV  ϰ0 pM6 .0xKKIb@A)ˈ"=Q|wy:](q0h(4L P"2ER# e!N$4̐l@@V*QTPD&DLp UUa0@҈1#HV&T4W9[ۺǘֽ'k϶pF wۚ]1+dtMyE.:y/'>`wRm9ޜ\\z>VF\[sř03k;'lhsMXj;|{+Ōr?b3[}IsDgv* [gI#)(2&p[?=%X @Lޣ\n<# bchAZL (.R@Da0MeEIVvݔ%/l"'GOD;^*DsLDs Hۍl %!BHC`   D%*qH"bZFDn9^4YB~s AT"vܺ`+> -MKm CkXb1ԁnk?圵:ƻ,HpS/r5ZqN|v:f}dVkF^dy93Poz1B®wчc\m]@ѼcOZxLT X}Q8 Df˅*5T@&(񎧻},txvgdNoKE-] `Sz#3i4`Շ>_|=4 Wa{ވ`ASհ8!*CtXF\OqgɟÚQ aGB0z0 (_vb>,.)E#DBYI (G:݄c")/zZͭgJ".Y0fqLĉDj4d=u+&\*lJ`y|0Gl%Ccr"6Kl_cNv#"-:xw7\~zS/(mŃ-_KF5GcA%6]u͏i˸I[UΣn #p~P"MJ{BBaF B48'3ξ`P7ʑ#AQкƏ#1@Lp vW_DA !HDGd[ә]b(ǤnKE|r(ìչ )tRkFg+=m+fYp-vJOJe`"q1zjK,eu_]KҎzݒ^=|]z% F#TjI\걿cB9pO}kߓ-.8bݥ}vkåjBe9 4)P̐󋝷;\v/{ lO#>w#uJ[Rm~qآoӗ7NoȟH[{+W'OMEo^P<](eCe# #B-(49 2Uj:Bd1&4,G_j>5n[q]-BNum5gX}֧7e;%SׄKZ+ORidE>GٖR1WH Y!fMZMKA BJYEKq il;#W[ȤGR";ҩ#EL5 ~)UCϋm`@5|*TʞnW6kx*ɉ閿lL<Seds+u9\(x eE(hd 6Fv?x @}o+R#- gZ tVx$}j_T rnBn+ ?e)(1F@1SDຏhOQ!kQ9*5ݿW(F<h\nf{N}%dμ r[J}b$7P:jxTvom'U F)T E-fvD*06UVJBEϏ6C/6]}s!y]/YvC`/oN{@Aɹ :~:mv9Q2K$AÓb_2L_oE}->.gn:)nN{oW_ku{+n../C8qI/Ҁl el8.oîAoMm&g˩9Bpt#d*fjpf.`:jS':Okqڒ~hpf%fZ/F/0qEf6Z0_Eѵ`&W\Ŕǂ@1j4(82BFX"၀%cI7%Zs%Z4Ohgm~oF4]9MB )Jq)T#DAN.5:UO ?wV 146ɬ/QŸjHhs3}/Y>^qQܭ\$p.n5wπ}j~JW[~|Z /OW}䢡l|שa3Z 뺤j;]y{Lnի}ٮ\ߝd!DSl ]dлbb:߈n^^JS$[M4ɦg x7! O'nN7btSѼ[zMwB޹v)&f#i}zEoN[Kp_E&Eg/?\=\䗛O_yЙ+L4R_&$SNR+NaV7Ԝ4 +f|M >mY)b.nWPsK৯7ߟx,Vc^AS`3BFe7vΑȳJ%$ LCأFԳ%0I) ^ I?Pczp8"u ÎזW̛,57qUOJ(پʈp5*f.{\rc } Qԁou"6#w' [aHgv.a|>|?$f=(qn2M \ Pɩ7遗]z 0őbFYAHfº?:I47HPh %cPBCa<\ PTzU*ҏd4HibCSU=ך3RisN{YeҨs*hF  mdIY"93̠*429Ǣ*di5# X;EiaPsKp ̩%S8ֻD-A2m* 2cUfLXn h~Ș{]gYnL>[2㶬Y#c1v?Ed&b(BU-?""z#)B7["je \:L &"vw0dnBLEFB "">HYNpv0yv#ѝn>LIw\3YI9&Sa}F0T C)4F[xunxN;)tZ/FOL1UR 1:b Пqc$Wqq`ШFA@fBЎK$ wIB/ge}8PN(C#H vك|G6#N#bMvw>Xd2n"VQVYjɹlL/nK  )R1IF t)o\] M9RNa!DSl ~n|dȉ1&~# yn d=X;7$2jw3izT BL'1m]` C{@ևslS$.4(]hЅ>W"am50uNX>eXԡjfޔVBk5|VzzV*B?}rVz\Ԇslj[ Rdd0CZƖŇ/}q\떰u|7&u7qG%-T*9FͧPI!QIjj烍UdlJ3ոXN'u jd܏sp=<_>S% .L̅\\RgRBRjC5jB:T&T$-FgOh x+R4k #*vK-q -䝛h˦k|yb726mrڊ@B_Q 4][\tFWZF۪s)@YX t,J, :C}^AtPv-N.-3[d;G..sdP\%9 ×JYN ?ϺA?թoxi Sk?}UWq`uf?iu]Ӈ/km:/MvX'U-Vr|ӣUkoE~W1d̮{w'C=oq[o2?|`N3捰?ǧ;|C5ۺ^@|Bsu6 ^:v>~6@{6i Nm)iWP2XĚȍ{_NkubDNrڋ?Ǘ`{Mm[>'l6if(f3Rc  4% 2|7I7o`JuIKͥ-]6\lH@e\4lUdi<%jDm tłsK$1@c[~VvqIےldЯIحa!D5SxXfJxJI4hȦV˾N-JIn63L캈5h&jie|LCS,/EmyQJ.`aŁo|+ߏa=#R mUNQBYRڇU t RAThN$ӑ}M Qrnc<Ď5漖kE0Ʀ޼!]z#Ä'9{@mHUC6 %I72PXd2Mf8 }B 'VJ ɔXX#L3VZҠ0RXbeAgLWZrrD+s2(+DKHQt,e".VV2ϹiL9!D}zf}= e )hp 7sϋm`6mZp(98nZ*{]}Կ/~%&Kg\7.mK3$TxnjFHD^B3eΕuVrB)9Ð%׊+hT6y-@`d0o`2YfHDU% QQL5;4PT:e)pOI]qĜwuH8{*06 ,; ̗Dbl/ٺ[}*6u$ZSWC LT'JP(@c٣%(\c!E3,Aq?2r ʐtK+A^m$֌PʳpwnNY2ɭwG_{4Kh8Z)jP7Q[b0< dع?ܛ߯c9Q2l'$~uF>W~BpwF|>L_a@ W80W+J[z.JFJ1A\ uxX!<|N\-kHDs_BV0]i KUw wd+i* W͙[EJoeߗW$턋wd’ X <,RT٨5 &,oڗ`=N=}íB1C9E8Fz,3FT 8IFLto} &ny6C9E8W~At{py薊:߈ns"to1 yp09etj;s5v8o8Zc1u7^XWUc dYKV |ȒX !<^ikӭ] IXµdBs /Y$2X"]&JVDZKV$LɊG|{  Z59/&ga=P\U$G!pODɠu)R4ZYR(iu2|wV5`_bFc)m;ba,=gc5 W^6KTPp,=gk랙,4 7υv,=gj+K/m \KWj-mÕR)<雵>0Ky$,hJT6crJ䱴zp*f,UwET1cicQ +K/JmSh~ u6R\R K?>'KYZm+K/hx,f4-hx,m3YʒFnjV\oqW}{#JeKleXe:SJ7-$h1n9R>)rN7KZ5rə.Q0Kliey#^ꔦi adM06b5W @Zn27y)SUɑpPԆ &z\Kne!$sPNkN/PX;S¯e97.YVFZvzc ^,p8Hժ;LUO_džX/xaHaI 6.MPڹofRn͏Rq89!ʕ`kEue^spԚ 9*W'Cز.FFj~?W?uI$J봄d~x 0 !jWP[;Ӛr)Rj2Ob`7eUW*(jtF[S!̳$lIY3\*cje^e+J! c*J=Ī0vc~PըMZ 9uI!`}/|x `oٛSMH\^+B˛;W}Rӭ~Pq җp_4{Njgu? Q߼@z#p \~Hs}FS v_oGD`jQn?~2v}~Bo=xj҂C?{v}nQ}$(\I==x^J['@ ag rgwov4j5G7s1yHck1=)s˴%dڰ7)=WGo31Zjx 0+n4pH-ǿ NX9%>/ŚY*Ci S5D49OOfTen*6X8}:/Q8=N;(NdiШ $@PdħhShq_);q'йn!_~MDCkG||jJ ;jhzb57$mwI .>pZ1`ULFTQzBn)ĩM[*~#&8Z.s$m8eR'nKP ^; M8t 8^r\:)hԜdFP+& OSjO|f$UAc4@nq7Jt9ƏqF,2ǍYY0r'al]y|QݥtW`N^t)hhsd Gb"15EF K~~!,h93^i203^;gBDYfx|b.QڀbȢ3/#5(]kS!VLCBUa,@;TVTU6TdХʔP؂( KRYVگVOUF=7X=*I}N`nIACT9U8Ri5$RsoW*uqAOWW#(~^]wA˯U}9g3i&S|"vFey^BU gLiU kkm,2KQ\Fo#F%4s䆟QГH21[C<ި˲T^YƑy+ckij2e*#qr*%3ĭ$ cKQ~ВD#(G)$s:)Q6ȃBk7Y*cxJw~>给9x9Z)PBޣ/57wǫ͉ :EO> tGhV@2'R峰 sV&f]v_ K IC[G #sl>: -c<y dR3G,rzy"iGq˃QL'j/zx"Ho$|t-{^'fPٷ^ls@Tq-U":bzqHM-ɔd%V .O'!3*}_WUjk~O7JBU㸫E*۱uZ]dav,S+d谙4VYU(׸4%r?`9J|1zP+K`f5g~X`FV[j5Dɭ ]"-b+^S ZuO\J+dȤA+ Da/ Ԗ YI)^{T/?0ǮUYcD)EHhCV; XJJV;JaɌ+8Q1aT[Ӏ1Me*Y+J J jTZ`d*j*t MsVV;Ik֤Q Jo ]=Л=OQ '0%4:+=?mlUane| /rtdiKUƒ@K?a㰢3Qi 7\03u@nu穃8uxCGbY=% /'; ۴4b0VbfgzEsoF F:N^N@z &mZK(;ZEȤbѠHFL|#S𩵓ɤρ2)C0S4S%ԑ )KŀCoDg_ާZcFT4ft!ϜIjm,uyݔEH%bKF݌!&K•!Ϝ9 [ZhKRRBsB7bTu:<-zw`kK*KQ5!Y"Xz]jczY*5R7ZF,=gkAYʻg ,mG9Y J f)<p_A}uAG`,=߂]A)Hb~2L#TYɪsB# % 2 3:|n^j|M9ܥ97ؾneNh!uyj΋/t7SHh> *>N4H+AZ%(r8i&0FH%t_޾t_I"[jSBm6IU)ң9/l]|a{R =yݟDK,;JՈe^ZSCVQ8$S?{㸍ÿvh\ &6hH"/)ɶ,6e=,w{mY*V}*ɪW4A9VQ+$$$8Pe!P~Gjv,R(iOEsx4@ q 9Ȳ~ DŊH-PZ^QvY?ϫuUΪz^ez!NqղFA,o-Y-F+,5-JA-zkz7-])! ʓIpVP:ucrX:'v_W0 }EHT$ԓ&7 2K5 [31Nt9.ӝ oCA ̆XscJ@8BoI%DK}PYb%*1<\;X4R`Tu 6!*l\g%%&GL#) "!MwtF)9l&_=lpVWR}HLvNLϓGr7&kbr'r_~~?o6_N-&wvU=ocHsA&ǣi8_|{/㒗 %OAGϹ?/gH۩|?x[PL$5>z4n=!@ "@zyӬ?wqUxe|k©7 IB {D/i/'դԇH_1Sh:'go!Yi5ROQ2XV^|?9oQK/IZҫB2jsDUDܮʛnʇ!_t~* Gތc&'H[ˋwlAޑm3rl*^jü0YBvfne\s1D5cJ{t99Uo.a/抜o.py!vԴΞ@QuۂmT KW`5XW9KPtb\f链7RvuX-ճV Z:5>]9YjWN|t)1'pHa©A+urwJ`5~Gj6fvĤȬSw«mxb\1jt|7GylS1L VMuj#Ky?|y*?4}Ɔc\?y' XMRf5Y,g&5Z6LyWS(η[Bh#L|VnMuϨ?:ܛwk@h#L5ɮw : Q˻oӴ޼[q_B8D`JbvRU:[O_ݾ] *lYyg[-RXs%V2E=2Ş@rL m4]=HqJXI'3= {aWr7orZJy%wkܭ-(=)vOz}5T&X0"iS:Mn$;zڿKژ~W ЯvSyh?r@tNs*f0dTw%yc+d[䴐(O3z؛'UK +`.%NJ][ 1%̅8ԿO3W?Bs;}|JD/!<o/_ߎ4uo05כh9Ϗf7````o`]MSlI]n.Y>9ܮP-]o4+/g{82*=Kr(WcJطqI86162D RhChjנǡ!D hB䖃"AץޅboWWɌ6B+VW~Q9b+_W_y-׭#ޗe7/bn hbYM"e+q6 ϶-'ãyde$zӗb_SJ3 $J$\#ŚIxG>]R&72 4J;h 2ä⸐}cPdGm[J,djAe$械~寋0!12R&JFEB5$Ua5G{w#%ڟY/S4{h3^g %D 0&²DAëլ lpuam#_` 8 yV`*_^ ~_;PLB{;vlSwc@t9 % s/eBh&XX8'Fkf4MQ+dWΌ cQ(d00~mW2y(+USmDw 1 T1iu$Y'Ec*LLi HRX_FM23VC)]g˯9IQ`M.W3{}?+S9'y+x~3U}6n+^=kaGtm6 rWnj1Ytv-ʀ5SٟSY~#<5ѧB&WH/aPJPtA):qZugqZm =CAcCv*}Lc7Z [6 $'nL7>>]j%9AcM7N䥨ؐtܲӥ//Is7Fi ow%Q#?r{6ZwvIZf/]t7{ql7rm\mB4ցkR: h}232݀n!|1I֒=hg4JJ7MCff!T y T>-G-!Cgy[S@3xE_Φ@{n&=[ym)qz7U8:wk4}FvU/vk^@h S),IW|ȶٳd:~jBy[:/Q/ⷻٻ$z?Rڞ?4x9s x~KrW}$DW1R"PJ2D_6Jg9K3@)#O+J/BK;QJs1J-0_8J=RRh ']8N9p4V &Iic_5&$}f Y(kBI ;2!B&gyO=NJ'1ؿH%jc+d 2Q!Qz*tTd%^;VX4ѡp:a)\IIY(i4BL+bp)]!!„qj7$*cbNw2@¯C:1FDr"fc!O%!$CQXK45*O5*FE9Sk:N 8 #u1:3Bl8Lb-hhX%$"RA(!C`pl 4QIo[EL.%gV%zR^Nޚ,J(E2 í !? /*"JS( sGY TvEjP/ᑪ IdC }颢vKeDMżO ?epKI8{׶ɍdEj y. {/;F#o-.aVL此r[`0HFz,=q/J&P^BRؐ{* d#NFz@+9M^~1v%z99OѴ9h$ 1#1465]hxWqװ-CjGɪo/5+-EUZK-ԩGF-Ö̓ J <1dzS+q0- y&dSt$[ wK tRF[RwB޸&GލpDUN7ۈ$; _zFPwB޸)#IRkay)J6!_x7s&ӏ0-U~0lA*%wp Xu zxnqbɌupkY;-\褃[FUj_bK8 ƒW1=bAhouհl0Wp_ۀ%12$`1#V9.B^顗VFBP:S/N/Q/a'$1klTYΙrW<G:{X䙭~qt?Ձ]?:鯳?`5@k>6Z>n*˩O0>/h~ӊvʏ;702̠䂂U[1|R),a!i[I1T| Gp+ T\( 'Lĉ$ _Hm4'ozg>\"Y?8u`누I|G-4ziiXInHORؔ yZyӋukR*<3L̇E ZOb ¬TbOj$%?lJm JJìpQb.f V`m:06LJu`R';֊e#T (rØg(]1LrRPFdT*\ ts\?7mV}7{]3~YZKjťrH5}ĦZ@QˮNUqJi퍠!PIr,)Z%Z. 7ͶFz<{ɝx/|jmը,"ó5 /ե1xd;&gYٞo ?NhAj[{tQ4z[SFکz@i`[J,J7Xh=:y@mhcm!p0G#Ԩpa$8H ]N>@̔u .%s:cBgNBqaG jv֯T;3  ܾ9*܃q&X[ S)3rFQVRD2V*CU'y)0T2ҴU@f"3K+s+lQ핉s%YyUdΨ|M fn|Yb8v_ֹ q <7[jLg*u. V ;'rCRPT*C {cCuwc,EG;c- mn#cMol"++_k;p"m֢_&1R4X Sre]ϒvکL+a8V(ep;A%LAy(e >FrxN036J ;% ׶ aǪI6)\G8"-i6's aԊ62(\ µg p]>=w kFhs c,,u.cEЉ$^H~gXµmқA[ꁯ7 w;΍aW77Q+kH U'뽽HZ6Mm/5 )[26&g$Hf plw/T4 +\]HLg*!Ro@nYfxC`C.,䍛hMn]QMIxT BL'1m@n[õLO n]X7$"F, )xT BL'1m@n[Qneޭ y&ڦ0q맰Z?ag' |\ :]svh]w龪iW$/=>e^ڛ .?Wu62=1%8yy~Fw@w6˚<9[1.{]~gn_Ok|L9z;| )Oլu 7͞,'5(߈ WB3/ܝ+>GgW*~/ي2]Q3󊽛I}1w|E쁝)'hb{CءUAeJ(*i?*!;$rW1 ~y6Gv]~gE&a`y)W;E>(βC]&̷3Z|fyt ᜿;2H^˄Uq'Q'(4_֏;aA(S#ȗjFjkFQόv d*e(Ug¸ٞo<. fj!k;ip^J;ipM74G2-imů\mh`\'?١ڬSԓN#)I5i uFns: nЅD{ :R+<\Ҕ1 p0']8vS`eA9s` '9*W*E"/I+di¡"S ]V7 B=Xٖ )-GKH WUyFdQ['}`'2Yݪ0δ31%JU9 VͅA8|z;><)22Bf/ Qw U-!`~!7}d$kO=RdET!XKO?@9% 〤F?h-FFe]j\jp%-TgB& G#=ՐttW6hd"Ӄw5+Ea,Dg_#ѥ Ixw &J}g:t88=gMJQep$afZP"Jrk!Jrs . *UAQbp YF'SUPUA4\=i#KbB#\\Z![FZaZ/YV:Ne" +$ Q$[c7EehDAU?:1'qWc߈vU՚ϳ/WWuxQ{=Gl+&%vH6$Dk] \F#}N0Y)7P +@tJ7+ʇL NBfNb@ ,YWY*Wł$͉6cx\f V~] [IS8Q)dJw"Lr$Vx%GZߣ&WTx3vBaNAGfaE ,oM1pj.;*K b3g2ʕ@m 5Lu bE hwJQF*;8]Q_|' [5Si@I!:.):!<v]y@qZ@a t653tIjC)z.d/$Ħ6BE PSbY%r J Ό%,,%q(e{mEs ݌$G Z%[W!&d,'kȵХDɿM9kJ%HQzRI(*ӈ |Vxޫ/[=uqU\fWOZV.{lgb} ͅlU(hCdeBiE+F܂$CU^r^[Pm)hO<=M~z۹q <Fb.:DpAV3((sRʨBdRJK,!U]].EM{]~\7kZOk_}q=?|grWK X7oz/Pu?}^l9_V:כRH 9kFK8_{r{wRǷ]Xb gZf ƞ]ewQ/̉^Kzϵ -]'z`Xg)x>HCʿKMKxJfAwtV6+ K\D r!ٗN ֝=7Ԏ__ =ۣnYd3n ٜ Z6=0LuH$x=| 5Vȟ}!CH-?]ArNS5&~ ~hH\ԽvkR'ͶZsf0>n®o 8׼SY45`5AM[_-%5 "iH-\E*Jd||aS:i =S1i9j6QQޅqM&ATލR11oxQ[z@wB޸&ٔ?&nlW1wK tRF<:8wKO? n]X76rb'Cg7|aSϊd4 ƑƁ69Wp B d$5vRV [{3yZCް. |ٮg?1U/?5(?/zzYZ+ttL2תD=F4۱1 jTĶZhtlx(6ֿj5'Ӗ=4F]$ fjBQxuQynelg腲}d>ݖۤnl r/VA]}fXuc_Y iZSIx^}%ka=Z/I >7N6ZϹcFO6iX7.S &ےcp[)n'*Re*2rGxjvDԀjSlyUGժʿ5[ޗHAvd\IBj8\;jkiJ ީX$x۷xwfSz.R@fB$@9(!%S KK4(@,ZfB*%w) ( .>wDݷ\"hrc%B.o0z %U:&Q}t`hDA5%R PNxڟ%G[j6%9}8$9ԣKֲ$NP`%>1һg4D&s@"$(8ozr&\K_ NAw3ڮBNjG+[dY fid:\s~<=?CzO]./6zХ[MvH鲖^i4Zz ?^⨄ 99zDMhkzQ|Rx4X4|ųzg\lQtqsсjB j. Lpwv@ ^ W76_w9ZQHێ_ow,Pou&ƫǯϟW&$YA* J~/mIIҺ3WiďC?٘|%dxB8Rr"L1ͤs4yASBI+44WCZUa#! u ME=3hrA1*;,"K-P0G8Y-Sr2B+|%ؚlg`V%غ91eEهbj6wF/F;,Je(Vy8.ǔa>*c`:hTәKx9.w^ɣí1.f~}؏N5DJ8#tŸq ۹v`(o. Q+ 0(3m"0ղQJ@f*ĔQ#}V=_zVh#'%$>iE$$": |+B\!4 \.=&0ec6AˍBl߭>3=P } t6F:V.XvnvK BWH[#pÓuL غ{F:r)ڳY0]R-NS.22E-6994?+Q 6ḻxg4HpZ[O{G!_4j)BLttÝ:CD.ĺm($eW18 ﰚ=-"B') oi $gEe g z7XC"a~kNs=c͛3lot nX1r%4ザ1XxHnA XRBY~<1O&n ZE!H *܅N=mHhZng,[ I(bdhým{CP9 G@KzפS6[FD0Ct4y0kFgkյ}>*Q($Ye9S+ƢQfǮNcn/@J PKމ"1^:>i`s5x[m>YsiUj#Xc# L)yon ¹(o?NXN +@õg|;\=Wij_ϸ6`[f]RbV`L1EX<, S2.ܱ*8sEA7p`Nf5 H>iW-zGlBع5$E*)dK˸%#=na`t8`:;~x!<1G:siܟ͈+0iub*O\HcOaY)jl]!J+AdNjxS2+: ۟ˆk")ޱAA.peWlF+lnI_ƙ-9d("kbKnآ3g͛iisú:Wֺ6FR<#!KճL+p~\PP$Ւw1JɏG1r' gDT)뛏ws9I`k D+1-ENt]8EX7|l~h}>%wZuUWOzۡ צ`oC/n>Rּ# dAN&))ԑ7y^߱{\YAu{hXsB=B2p)yh‘ #DXtF]Y8sa)M_JѦ-,wE.u Ir^hPڍ;mI+;6*8OWL Jpڞ6 A-m"\d8DeD,4>1mwЎ6*^U<%M0 )e),lS8Mimb!AUOAI>):*]T ЦS s̓:~# l)d>Y ѺQsIVbRgTJlsxQ\_kXRgԄrsj2edaȁ Ca^EkCw|1\WSRTSok4M]2&_X1ƇeI35мkx"Z6}nRoSOC=yV  hKY1?i4Q D%2"Ju4Qxa|_ɆkI] :@w!Hove2aYПi*76/e V{" =$1^ʍ#5gʢ_A\g쳤(F,Dr*Z {/4j.^LvoǓIY*}66]eIF^S̏W3ҥR%vҸ.%AgX{r]^1" V)VIQ%u~hI9o;vyg[%yrYa#e\.C.3?_CRjzPi)` e*ϰ{S9o |m剈>}ϧQU|}4'yl\7le߰fwveus+ 3z?0@Q%̨fT 3&̺gJJޮsmL\rMU>ṿ_Ʒ>%oq]^Ez]= 5v=2u'8f'h !d rJ$>7qwF^j[!0 U|P4,fջF)xVk^ONRZ[w&k[&ݣM>މ=֚{S}%_=LB"z~e1`M<8%X"qOHđ"% JYu f+5`Qw&vQtԒ+mHf#2څ=()QH6n7bQd-QŬ""Ȍo/0 )rYѲ%n֕)h ˈB }|7M՜\sf}:R?mͽ5'P\VM>Y5G9}-EA-7bN|gwn/NƗ5+TTZ@”k;h$G*s^kvZc 4T\FRTRiE7:hoe* MBcC;2Qpuqv5ɩ0!,ql҄nW2rfK$dRH0!CC+2F %!^hNrI552,+o2}O#[|j KH"rBk fε vYMt50ݕHr%7-rA^ J3:Y.w HL,@ 9l\[W X'5gM76[1rǟO++:!.FCEv30$R#5ߛ".Qz\(ɥO!l5BOSK\77za3z R_\ps> *O$K@<BӉ; m T۽17cM TftK^])sְBo(moKq͍bL!OQXen=S*=GN:&"繐ڠ%Li3ebpv/ϼRXИ&h "@'P 0< (!:RW>W ]f! :1?-RuJK|cAעox=]BWpj(}Be抌Tyi>c.2qJ[R77Ȗ /JKMBV^KbO\ -]H9s8X_=Q3` 7 zs o=е Q39J :ƏkOm#*0u{Fs~UUH>$&R8O#0h*9`mG27ɴ7{M⠅ڸAz $et d"PRƑ,R[@G?L6^pP|nE):ze N舐TZKJ})< 2겒;U8QBnǐ'hp˾ٴ4]iXuKjѻ6R&ް+S ]7Ġ5]&(|)MjeGpI 粍Vwt̒q$NG99>}NX}S&mVNi²RZ);SvEe2Jꀃ/#4>[ʩL)A[1 o OL:j.Y mC2Z$2Ƅatt8`R*GET :mHR*UTs󞹬Si–L*h#Ujۑ@R ZbҐwFe2hiOdK.Ti}FB*2Nc\`d)ۅZvT9p;iQwQ?r xY?/ƹ|g o?~{Eh&7?'~gUʢqe9C"WKۻ#T*tɑ_fjSj.nݹ ӵ`M GL,D64_xW\7TS{.`c`+j gR Î:󴕟7. p.Wq0`hˉ-Gehl#|?2Z y>jKެi#04 ۙ.ķ5G[2+hx洹t|o׹#n'^Hl/zl}Y\W27ULC O2ITr+l&Wi5DII$TG=gLҬ\rF/;&BJ-([AO,Nx]unT [+6I6@nqT72?aAp$$*ݹCoɍ '$Hg4^Q &ݮ%eS 6{/Ώv~N{ >b]O-UV"/?\\?MJe$Aߊ1=9W'+i6 .-SL `Karbeާp݈JjJav&Γ{#`;r_ʖ$yi+m[lG×Dvm'mICxάeֺVoi c}b|wJC?5}TY0c뒓Y\r W2&<^vF7M(ⴠ}An0#UE[4(W~|߹9#;3/+a?Bh2=Px̬e/깗MrkqȻڒϏЂϝ}#4rqےEt'-7M0*AJ&;n}[~1vkCB"Li} 9+ ήF=g!d̾wIj6$O.)o+"KqzOHlGgN9;2vW/op| x!bd!i/H$ =1..}%IFDV@bxJWLdwƫ] `+ 7:jٚJ5_2Ry""~):⣢Zg,=gf եDM%` z2σRHChй:Ƭ 54 ۘby9/YkV;3}3}y0ϴt{mcp(7͓ UJQYqu,![ҤQ/- -v,UOW9, rHK8EL&"R|uWY,}zaB F$j$;4:Y[ Ib)hQr߹ rܾ$.7~ܺ*k2:XeL3M,G^e-`ٰၜ0.ˆ 67;a^@J7du֖U/[Jl&Lvܶ[`=1˪%GY>Y"axq9`@s4klTstkƭ05{cra&7n~+,'_?35Cf(-,XEj]u)_g[Q4vҁXP28HVydW\95N!{}p`o f"ּ:nbB^Bw[xmRB_*-(ק04EѼΰC12*+UJKkj+hFnMw)P8]a|hͪc L_76F gbggR ~ (/A 턷B.))nB2{% 8DdA`)kdBzyyei0&Na )kTi͠<Hd?; N-Ryo-`ӬOVg.۩wmmJ6g и4ZoTݚ9'd^vT7K2d((+#Sd׍Fh|}9@!xkZ'r \ h6{B?\c# J J܇uXmѾMDb{5>JlH2K@5(mڝQTKem{fiS:pKv1UNPzvWq??KQU&ȘE`6_-nyscGwm9DӷG tL͉Fu6狇\"ŭu%lߏ9"C n tj$p%$q (x^޺W++͑e,P9x}[bJK<8ԗ9Z99^ -wq;=SMQf*!5 ]ߜeb4̍fư\$gR jp&a?M'I#2Q5$UNR3@N@7{h+`dAGmg $P xb; r򈱲WPGijJXhVaQ. u5 -+(1Tw2)(4<UT':nj|!>3UQtωiH'E.y":h*~C.BRf U(˄DbYhR#KnW.֩Wi?Nd¸̵rS/A9gZoYdKS19^]jBخL /CUI4vae:Rהy4 X=Q6q?KwVO멙 F⬕3ፄi JI7] \hE5K;xtgz$\K.>ڵz xAT˟.-=]pjL78b.NJŖcȏܻmO S%-6un.ؤnw˛U ˔ؔ*Ϋ?=/*>}JrйyvgG6%/3?nzlZPYªأX4:[rS{ɋM ;I@xT=m,J\m!/E)n-l3Ԋ[(>&ѭTJ I<[]4ȧ3έF^[ ^FѭUih-yotkC^V|ʓ;6dj}!;1nvU^/mQѫ*R/ʣ< xNˊԜt/eUj$DRD3ӧRBj)K'/=F/Rp3s}z):]KKNȸ)>'eUj'/=j/y(RȌ/<ҕxZ=9zs祾JtTy)usZKBjS^z^S]m˃RFZ{R]^Bj s%YP |>Y‵&VjC^cVtB!DP |\'M[󬐴{b8h$繓^x.<EĮsR_VOQǼ+ΩW8y/R#|L`gwlU 0o ̗33ϗE}GC8~?o~m~>^Ϳv٨. 츴*,{W|4 ;ċHq^]mߛ5f)Umb z0JsMK}fZ0s!v4]]~(Pm dR*Ԙ?VqaLjhArY˫,l;P*XއuTZj[FV[7%ؑbGʹwo~h~WE$[u-S̮n3_a*KjQ`i%#nrQNV&kXU"EEjރHvgj"|22ΎgZ4ݫLq.7CUq~f/+YEi.`dStjf Tr_'@Kȏ}})*y9-~`i[i>\ Jr m-B[x +$+!殦j,S@8 ΩD#'( N)AfuܔRʩ2[C9 Bcs[jQp\7_L/ooOf(/x]ELǔBpr&V,U*Ic2EцV=eݳDlDFD00 E;E3.Z! l"qdLn?j$Ld)!$Dp&ڂ V!B07뇋qy?&:q%ob+lUz-W녽P_)@1Ņx%\yL:(ŝ^L_o]5řʙW23Rf*5Rw QeӷG r3&/A゚ kwsihs5өS~Κ J/cbS\ ]F_EzkxkoʹҌae [H5{mxx ek~Cv^JfnSUyrf׹+;f}>z*5"LMx< wvnc1Z ~ou௧!8گ Bq5 ;)ΞX3MSkf3w1== BBP쎈q&b\]ݹhY =-v yHuL  @Fi|?8R0"t%9ګu>ΫG9-׹;ЌQݣEgQs՝RcՊ_'AF@4PCnL ޲ Zk^_6VQ-v8}E+~HB.R3 .MMKԡO:6<䅻hOi19z;D- }F_lZʻk=[]4ħg[эaDP |\'M[N,nmx wѽ}J ZSnH9*l;Qa[g[ѽL4p /GY6*W{ޱ1@ұΕ1"1K csz0JsMK}>&TaT`rlom(+Ct <ɯ". Km:'$NUy GHI("g s; Ϩ9lX*yy/Cee2nҙ4 W{g6>E^?{u;)4=^ŖcRԗNd5LsIs@#cݰ5qKeXPo/Uv}r܍*B@gC2,Y.)9&6wi(h)>u]RjLЇ&*qM5-5i'jk):>1=SYUZ$..gyEVWJ3pb86؈ 1C4J7c["+'*&l72NUowCM;]c'|%!g~D'I.===>kvJ~;{M#+䒗.h|[DdF dp/e4h[Pak Gj6-5t| F5c۪X#!)`Ի$$,CǷHDVvHj)Ռ抿HwOؾIE?;4 h 蝦G&e) Qz 2cDr.ZS[T 49[nN3rۊO#/%F)<]ȧxItӺ-gVA>#'XM{_ JnSxȍ_)var/home/core/zuul-output/logs/kubelet.log0000644000000000000000004662272615144066624017724 0ustar rootrootFeb 14 11:10:08 crc systemd[1]: Starting Kubernetes Kubelet... Feb 14 11:10:08 crc restorecon[4570]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:08 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 14 11:10:09 crc restorecon[4570]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 14 11:10:09 crc restorecon[4570]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 14 11:10:09 crc kubenswrapper[4904]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 14 11:10:09 crc kubenswrapper[4904]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 14 11:10:09 crc kubenswrapper[4904]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 14 11:10:09 crc kubenswrapper[4904]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 14 11:10:09 crc kubenswrapper[4904]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 14 11:10:09 crc kubenswrapper[4904]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.643446 4904 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649721 4904 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649753 4904 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649764 4904 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649773 4904 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649781 4904 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649789 4904 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649799 4904 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649807 4904 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649816 4904 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649825 4904 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649860 4904 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649869 4904 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649877 4904 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649886 4904 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649894 4904 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649902 4904 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649910 4904 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649917 4904 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649925 4904 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649933 4904 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649940 4904 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649948 4904 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649955 4904 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649963 4904 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649971 4904 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649978 4904 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649986 4904 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.649993 4904 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650002 4904 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650009 4904 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650037 4904 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650046 4904 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650053 4904 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650061 4904 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650070 4904 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650080 4904 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650091 4904 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650100 4904 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650110 4904 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650119 4904 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650128 4904 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650136 4904 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650144 4904 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650151 4904 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650159 4904 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650167 4904 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650179 4904 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650192 4904 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650204 4904 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650215 4904 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650226 4904 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650237 4904 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650247 4904 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650257 4904 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650271 4904 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650284 4904 feature_gate.go:330] unrecognized feature gate: Example Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650292 4904 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650300 4904 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650309 4904 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650317 4904 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650325 4904 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650335 4904 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650343 4904 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650350 4904 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650360 4904 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650369 4904 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650379 4904 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650392 4904 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650404 4904 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650414 4904 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.650424 4904 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650655 4904 flags.go:64] FLAG: --address="0.0.0.0" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650675 4904 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650698 4904 flags.go:64] FLAG: --anonymous-auth="true" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650709 4904 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650722 4904 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650732 4904 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650744 4904 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650766 4904 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650776 4904 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650785 4904 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650795 4904 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650864 4904 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650873 4904 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650883 4904 flags.go:64] FLAG: --cgroup-root="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650891 4904 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650901 4904 flags.go:64] FLAG: --client-ca-file="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650910 4904 flags.go:64] FLAG: --cloud-config="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650918 4904 flags.go:64] FLAG: --cloud-provider="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650927 4904 flags.go:64] FLAG: --cluster-dns="[]" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650944 4904 flags.go:64] FLAG: --cluster-domain="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650954 4904 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650963 4904 flags.go:64] FLAG: --config-dir="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650972 4904 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650981 4904 flags.go:64] FLAG: --container-log-max-files="5" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.650992 4904 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651001 4904 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651011 4904 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651020 4904 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651030 4904 flags.go:64] FLAG: --contention-profiling="false" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651038 4904 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651048 4904 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651057 4904 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651066 4904 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651077 4904 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651086 4904 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651112 4904 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651121 4904 flags.go:64] FLAG: --enable-load-reader="false" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651130 4904 flags.go:64] FLAG: --enable-server="true" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651139 4904 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651158 4904 flags.go:64] FLAG: --event-burst="100" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651167 4904 flags.go:64] FLAG: --event-qps="50" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651177 4904 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651186 4904 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651200 4904 flags.go:64] FLAG: --eviction-hard="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651212 4904 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651221 4904 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651230 4904 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651239 4904 flags.go:64] FLAG: --eviction-soft="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651248 4904 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651257 4904 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651266 4904 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651275 4904 flags.go:64] FLAG: --experimental-mounter-path="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651284 4904 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651293 4904 flags.go:64] FLAG: --fail-swap-on="true" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651301 4904 flags.go:64] FLAG: --feature-gates="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651312 4904 flags.go:64] FLAG: --file-check-frequency="20s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651321 4904 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651331 4904 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651341 4904 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651350 4904 flags.go:64] FLAG: --healthz-port="10248" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651359 4904 flags.go:64] FLAG: --help="false" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651368 4904 flags.go:64] FLAG: --hostname-override="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651378 4904 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651387 4904 flags.go:64] FLAG: --http-check-frequency="20s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651396 4904 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651405 4904 flags.go:64] FLAG: --image-credential-provider-config="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651413 4904 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651422 4904 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651431 4904 flags.go:64] FLAG: --image-service-endpoint="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651440 4904 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651449 4904 flags.go:64] FLAG: --kube-api-burst="100" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651459 4904 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651472 4904 flags.go:64] FLAG: --kube-api-qps="50" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651484 4904 flags.go:64] FLAG: --kube-reserved="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651495 4904 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651507 4904 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651518 4904 flags.go:64] FLAG: --kubelet-cgroups="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651527 4904 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651538 4904 flags.go:64] FLAG: --lock-file="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651551 4904 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651560 4904 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651569 4904 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651597 4904 flags.go:64] FLAG: --log-json-split-stream="false" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651607 4904 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651616 4904 flags.go:64] FLAG: --log-text-split-stream="false" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651625 4904 flags.go:64] FLAG: --logging-format="text" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651634 4904 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651643 4904 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651652 4904 flags.go:64] FLAG: --manifest-url="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651661 4904 flags.go:64] FLAG: --manifest-url-header="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651673 4904 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651682 4904 flags.go:64] FLAG: --max-open-files="1000000" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651693 4904 flags.go:64] FLAG: --max-pods="110" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651702 4904 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651711 4904 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651721 4904 flags.go:64] FLAG: --memory-manager-policy="None" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651729 4904 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651739 4904 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651748 4904 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651757 4904 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651776 4904 flags.go:64] FLAG: --node-status-max-images="50" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651785 4904 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651794 4904 flags.go:64] FLAG: --oom-score-adj="-999" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651803 4904 flags.go:64] FLAG: --pod-cidr="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651812 4904 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651826 4904 flags.go:64] FLAG: --pod-manifest-path="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651861 4904 flags.go:64] FLAG: --pod-max-pids="-1" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651870 4904 flags.go:64] FLAG: --pods-per-core="0" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651880 4904 flags.go:64] FLAG: --port="10250" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651888 4904 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651897 4904 flags.go:64] FLAG: --provider-id="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651906 4904 flags.go:64] FLAG: --qos-reserved="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651915 4904 flags.go:64] FLAG: --read-only-port="10255" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651924 4904 flags.go:64] FLAG: --register-node="true" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651933 4904 flags.go:64] FLAG: --register-schedulable="true" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651945 4904 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651960 4904 flags.go:64] FLAG: --registry-burst="10" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651968 4904 flags.go:64] FLAG: --registry-qps="5" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651977 4904 flags.go:64] FLAG: --reserved-cpus="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651986 4904 flags.go:64] FLAG: --reserved-memory="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.651997 4904 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652006 4904 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652015 4904 flags.go:64] FLAG: --rotate-certificates="false" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652024 4904 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652034 4904 flags.go:64] FLAG: --runonce="false" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652042 4904 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652052 4904 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652061 4904 flags.go:64] FLAG: --seccomp-default="false" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652070 4904 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652079 4904 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652088 4904 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652098 4904 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652107 4904 flags.go:64] FLAG: --storage-driver-password="root" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652116 4904 flags.go:64] FLAG: --storage-driver-secure="false" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652125 4904 flags.go:64] FLAG: --storage-driver-table="stats" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652134 4904 flags.go:64] FLAG: --storage-driver-user="root" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652142 4904 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652151 4904 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652160 4904 flags.go:64] FLAG: --system-cgroups="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652169 4904 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652183 4904 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652192 4904 flags.go:64] FLAG: --tls-cert-file="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652200 4904 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652217 4904 flags.go:64] FLAG: --tls-min-version="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652226 4904 flags.go:64] FLAG: --tls-private-key-file="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652235 4904 flags.go:64] FLAG: --topology-manager-policy="none" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652244 4904 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652252 4904 flags.go:64] FLAG: --topology-manager-scope="container" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652261 4904 flags.go:64] FLAG: --v="2" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652272 4904 flags.go:64] FLAG: --version="false" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652284 4904 flags.go:64] FLAG: --vmodule="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652296 4904 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.652306 4904 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.652565 4904 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.652590 4904 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.652605 4904 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.652615 4904 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.652628 4904 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.652638 4904 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.652649 4904 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.652660 4904 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.652671 4904 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.652682 4904 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.652692 4904 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.652701 4904 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.652710 4904 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.652720 4904 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.652730 4904 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.652740 4904 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.652750 4904 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.652760 4904 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.653946 4904 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.653973 4904 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.653986 4904 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.653999 4904 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654010 4904 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654021 4904 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654032 4904 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654044 4904 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654054 4904 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654064 4904 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654074 4904 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654083 4904 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654093 4904 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654103 4904 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654113 4904 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654124 4904 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654138 4904 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654151 4904 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654164 4904 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654176 4904 feature_gate.go:330] unrecognized feature gate: Example Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654188 4904 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654199 4904 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654209 4904 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654219 4904 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654229 4904 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654239 4904 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654249 4904 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654260 4904 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654270 4904 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654278 4904 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654289 4904 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654299 4904 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654307 4904 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654315 4904 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654324 4904 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654333 4904 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654342 4904 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654350 4904 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654358 4904 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654366 4904 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654374 4904 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654382 4904 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654390 4904 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654398 4904 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654405 4904 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654413 4904 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654423 4904 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654433 4904 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654441 4904 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654449 4904 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654460 4904 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654472 4904 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.654486 4904 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.654514 4904 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.663909 4904 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.663957 4904 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664043 4904 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664053 4904 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664059 4904 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664065 4904 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664070 4904 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664076 4904 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664083 4904 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664089 4904 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664094 4904 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664099 4904 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664104 4904 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664108 4904 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664113 4904 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664117 4904 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664121 4904 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664125 4904 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664129 4904 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664134 4904 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664138 4904 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664142 4904 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664148 4904 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664155 4904 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664160 4904 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664165 4904 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664169 4904 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664175 4904 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664179 4904 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664184 4904 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664189 4904 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664192 4904 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664196 4904 feature_gate.go:330] unrecognized feature gate: Example Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664201 4904 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664205 4904 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664209 4904 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664213 4904 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664217 4904 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664221 4904 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664226 4904 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664232 4904 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664238 4904 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664242 4904 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664247 4904 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664252 4904 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664256 4904 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664260 4904 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664265 4904 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664269 4904 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664273 4904 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664278 4904 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664283 4904 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664287 4904 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664292 4904 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664297 4904 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664301 4904 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664306 4904 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664310 4904 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664314 4904 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664318 4904 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664325 4904 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664331 4904 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664336 4904 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664342 4904 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664346 4904 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664350 4904 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664355 4904 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664359 4904 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664363 4904 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664367 4904 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664373 4904 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664378 4904 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664383 4904 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.664391 4904 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664529 4904 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664538 4904 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664545 4904 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664551 4904 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664556 4904 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664561 4904 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664566 4904 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664570 4904 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664575 4904 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664581 4904 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664586 4904 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664590 4904 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664595 4904 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664599 4904 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664604 4904 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664608 4904 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664612 4904 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664616 4904 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664621 4904 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664625 4904 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664629 4904 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664634 4904 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664639 4904 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664643 4904 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664648 4904 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664654 4904 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664660 4904 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664665 4904 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664669 4904 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664675 4904 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664681 4904 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664686 4904 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664691 4904 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664696 4904 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664703 4904 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664709 4904 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664713 4904 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664718 4904 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664723 4904 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664728 4904 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664734 4904 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664740 4904 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664746 4904 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664751 4904 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664756 4904 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664762 4904 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664769 4904 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664775 4904 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664781 4904 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664786 4904 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664791 4904 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664796 4904 feature_gate.go:330] unrecognized feature gate: Example Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664801 4904 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664808 4904 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664813 4904 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664818 4904 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664823 4904 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664828 4904 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664852 4904 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664858 4904 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664863 4904 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664869 4904 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664874 4904 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664878 4904 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664883 4904 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664887 4904 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664893 4904 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664897 4904 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664902 4904 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664907 4904 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.664911 4904 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.664920 4904 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.665107 4904 server.go:940] "Client rotation is on, will bootstrap in background" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.669764 4904 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.669894 4904 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.672148 4904 server.go:997] "Starting client certificate rotation" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.672183 4904 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.672969 4904 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-15 21:53:38.33546695 +0000 UTC Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.673062 4904 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.695889 4904 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 14 11:10:09 crc kubenswrapper[4904]: E0214 11:10:09.697957 4904 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.698737 4904 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.710921 4904 log.go:25] "Validated CRI v1 runtime API" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.739139 4904 log.go:25] "Validated CRI v1 image API" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.740892 4904 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.744868 4904 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-14-11-04-46-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.744891 4904 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.754961 4904 manager.go:217] Machine: {Timestamp:2026-02-14 11:10:09.752922352 +0000 UTC m=+0.565687033 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2799998 MemoryCapacity:25199472640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:8976ab10-78e1-4979-8878-7acc2b1073d8 BootID:6bfee4df-cb7e-45ab-a7f5-35741705b859 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039894528 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599738368 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076107 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599734272 Type:vfs Inodes:3076107 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:c6:ab:4f Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:c6:ab:4f Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:83:12:64 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:18:82:40 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:5f:e6:8f Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:1f:65:89 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:0e:00:64:3d:a7:13 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:32:ff:84:af:93:45 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199472640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.755126 4904 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.755224 4904 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.755426 4904 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.755555 4904 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.755596 4904 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.755747 4904 topology_manager.go:138] "Creating topology manager with none policy" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.755755 4904 container_manager_linux.go:303] "Creating device plugin manager" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.756207 4904 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.756234 4904 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.757106 4904 state_mem.go:36] "Initialized new in-memory state store" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.757447 4904 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.760948 4904 kubelet.go:418] "Attempting to sync node with API server" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.760967 4904 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.761004 4904 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.761015 4904 kubelet.go:324] "Adding apiserver pod source" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.761025 4904 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.764083 4904 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.764793 4904 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Feb 14 11:10:09 crc kubenswrapper[4904]: E0214 11:10:09.764873 4904 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.765212 4904 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.765225 4904 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Feb 14 11:10:09 crc kubenswrapper[4904]: E0214 11:10:09.765417 4904 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.767659 4904 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.768900 4904 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.768937 4904 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.768949 4904 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.768981 4904 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.769008 4904 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.769020 4904 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.769032 4904 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.769051 4904 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.769063 4904 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.769075 4904 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.769108 4904 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.769120 4904 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.769978 4904 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.770563 4904 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.770594 4904 server.go:1280] "Started kubelet" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.770898 4904 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.771900 4904 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.772492 4904 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.773945 4904 server.go:460] "Adding debug handlers to kubelet server" Feb 14 11:10:09 crc systemd[1]: Started Kubernetes Kubelet. Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.774321 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.774355 4904 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.774921 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 10:40:01.709862621 +0000 UTC Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.775099 4904 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.775115 4904 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.775228 4904 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 14 11:10:09 crc kubenswrapper[4904]: E0214 11:10:09.775226 4904 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 14 11:10:09 crc kubenswrapper[4904]: E0214 11:10:09.776393 4904 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="200ms" Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.777702 4904 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Feb 14 11:10:09 crc kubenswrapper[4904]: E0214 11:10:09.777775 4904 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.778134 4904 factory.go:153] Registering CRI-O factory Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.779044 4904 factory.go:221] Registration of the crio container factory successfully Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.779135 4904 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.779146 4904 factory.go:55] Registering systemd factory Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.779154 4904 factory.go:221] Registration of the systemd container factory successfully Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.779175 4904 factory.go:103] Registering Raw factory Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.779189 4904 manager.go:1196] Started watching for new ooms in manager Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.786236 4904 manager.go:319] Starting recovery of all containers Feb 14 11:10:09 crc kubenswrapper[4904]: E0214 11:10:09.789739 4904 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.17:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1894186cf8a545d7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-14 11:10:09.770563031 +0000 UTC m=+0.583327712,LastTimestamp:2026-02-14 11:10:09.770563031 +0000 UTC m=+0.583327712,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797432 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797486 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797502 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797516 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797529 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797540 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797553 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797565 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797579 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797594 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797609 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797622 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797635 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797650 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797665 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797676 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797688 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797701 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797713 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797727 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797739 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797771 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797795 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797811 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797825 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797852 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797868 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797880 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.797891 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798040 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798057 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798067 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798080 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798092 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798104 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798114 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798127 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798138 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798150 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798162 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798173 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798186 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798199 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798212 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798224 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798235 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798246 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798257 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798270 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798283 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798299 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798311 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798328 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798342 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798356 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798368 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798382 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798587 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798603 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798617 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798629 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798643 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798657 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798669 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798684 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798694 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798704 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798713 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798725 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798736 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798780 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798892 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798909 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798923 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798937 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798951 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798962 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798971 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798980 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.798991 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799000 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799011 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799021 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799032 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799041 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799053 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799063 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799073 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799083 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799092 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799103 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799114 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799126 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799138 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799154 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799167 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799190 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799200 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799211 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799221 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799234 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799243 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799253 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799265 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799279 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799289 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799300 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799311 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799320 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799329 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799342 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799352 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799360 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799370 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799383 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799394 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799406 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799417 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799428 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799439 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799448 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799456 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799465 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799474 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799483 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799491 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.799499 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801074 4904 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801134 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801159 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801169 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801178 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801187 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801198 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801206 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801216 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801225 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801235 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801244 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801253 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801264 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801274 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801283 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801292 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801301 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801310 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801318 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801328 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801337 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801346 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801355 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801363 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801374 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801385 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801397 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801407 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801415 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801435 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801444 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801452 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801462 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801470 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801480 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801488 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801496 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801505 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801514 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801523 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801531 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801540 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801548 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801556 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801565 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801573 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801582 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801590 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801600 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801612 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801620 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801631 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801641 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801652 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801661 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801670 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801679 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801688 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801696 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801704 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801713 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801723 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801731 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801739 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801747 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801756 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801764 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801772 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801780 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801788 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801796 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801807 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801825 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801849 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801857 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801865 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801874 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801884 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801892 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801901 4904 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801909 4904 reconstruct.go:97] "Volume reconstruction finished" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.801916 4904 reconciler.go:26] "Reconciler: start to sync state" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.817644 4904 manager.go:324] Recovery completed Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.832186 4904 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.833112 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.835085 4904 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.835112 4904 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.835131 4904 kubelet.go:2335] "Starting kubelet main sync loop" Feb 14 11:10:09 crc kubenswrapper[4904]: E0214 11:10:09.835167 4904 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.838785 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.838810 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.838819 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:09 crc kubenswrapper[4904]: W0214 11:10:09.838783 4904 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Feb 14 11:10:09 crc kubenswrapper[4904]: E0214 11:10:09.839127 4904 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.840252 4904 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.840276 4904 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.840296 4904 state_mem.go:36] "Initialized new in-memory state store" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.856660 4904 policy_none.go:49] "None policy: Start" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.857461 4904 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.857561 4904 state_mem.go:35] "Initializing new in-memory state store" Feb 14 11:10:09 crc kubenswrapper[4904]: E0214 11:10:09.875505 4904 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.902922 4904 manager.go:334] "Starting Device Plugin manager" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.902972 4904 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.902984 4904 server.go:79] "Starting device plugin registration server" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.903378 4904 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.903393 4904 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.903961 4904 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.904063 4904 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.904070 4904 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 14 11:10:09 crc kubenswrapper[4904]: E0214 11:10:09.914244 4904 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.935608 4904 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.935697 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.937038 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.937079 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.937089 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.937259 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.938006 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.938026 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.938034 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.938052 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.938076 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.938115 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.938286 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.938337 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.938875 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.938926 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.938962 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.938975 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.938940 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.939127 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.939173 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.939189 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.939199 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.939431 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.939551 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.939626 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.940473 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.940516 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.940529 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.940772 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.940830 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.940894 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.941098 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.941440 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.941477 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.942096 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.942166 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.942229 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.942262 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.942334 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.942345 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.942743 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.942827 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.943447 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.943551 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:09 crc kubenswrapper[4904]: I0214 11:10:09.943609 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:09 crc kubenswrapper[4904]: E0214 11:10:09.977061 4904 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="400ms" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.003337 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.003371 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.003391 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.003407 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.003457 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.003475 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.003500 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.003548 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.003569 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.003586 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.003606 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.003649 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.003718 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.003811 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.003898 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.003926 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.004478 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.004516 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.004546 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.004574 4904 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 14 11:10:10 crc kubenswrapper[4904]: E0214 11:10:10.004984 4904 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.17:6443: connect: connection refused" node="crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.105672 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.105738 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.105774 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.105803 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.105831 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.105883 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.105894 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.105908 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.105937 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.105987 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106071 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106072 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106095 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106105 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106096 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106127 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106127 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106169 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106211 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106240 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106262 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106282 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106296 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106301 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106353 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106370 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106389 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106341 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106426 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.106376 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.206031 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.208128 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.208176 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.208194 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.208252 4904 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 14 11:10:10 crc kubenswrapper[4904]: E0214 11:10:10.208690 4904 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.17:6443: connect: connection refused" node="crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.262234 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.276251 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.293575 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.299526 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.304801 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 14 11:10:10 crc kubenswrapper[4904]: W0214 11:10:10.314385 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-43fc2fe725b883e9e06518b3a9198c3198fa7f5a0a5ccf8136229d12c1210450 WatchSource:0}: Error finding container 43fc2fe725b883e9e06518b3a9198c3198fa7f5a0a5ccf8136229d12c1210450: Status 404 returned error can't find the container with id 43fc2fe725b883e9e06518b3a9198c3198fa7f5a0a5ccf8136229d12c1210450 Feb 14 11:10:10 crc kubenswrapper[4904]: W0214 11:10:10.315518 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-14dcae69a98f28901aae49b09877dce7a526f99d8c6229138b0d3672e0d318a6 WatchSource:0}: Error finding container 14dcae69a98f28901aae49b09877dce7a526f99d8c6229138b0d3672e0d318a6: Status 404 returned error can't find the container with id 14dcae69a98f28901aae49b09877dce7a526f99d8c6229138b0d3672e0d318a6 Feb 14 11:10:10 crc kubenswrapper[4904]: W0214 11:10:10.321375 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-2a671812bf72b40078117872711b366d6483b90d200a193539c9a1ef87be6c48 WatchSource:0}: Error finding container 2a671812bf72b40078117872711b366d6483b90d200a193539c9a1ef87be6c48: Status 404 returned error can't find the container with id 2a671812bf72b40078117872711b366d6483b90d200a193539c9a1ef87be6c48 Feb 14 11:10:10 crc kubenswrapper[4904]: W0214 11:10:10.327465 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-6cf3628a9172eed0ac602e5f77596d7a12a3ce7f2c0857fb13ceb66245060f44 WatchSource:0}: Error finding container 6cf3628a9172eed0ac602e5f77596d7a12a3ce7f2c0857fb13ceb66245060f44: Status 404 returned error can't find the container with id 6cf3628a9172eed0ac602e5f77596d7a12a3ce7f2c0857fb13ceb66245060f44 Feb 14 11:10:10 crc kubenswrapper[4904]: W0214 11:10:10.331538 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-9edaf241e93f31ded2eba6b7e6c78b76da34b442735a5c27ec5f7f9fcfb99197 WatchSource:0}: Error finding container 9edaf241e93f31ded2eba6b7e6c78b76da34b442735a5c27ec5f7f9fcfb99197: Status 404 returned error can't find the container with id 9edaf241e93f31ded2eba6b7e6c78b76da34b442735a5c27ec5f7f9fcfb99197 Feb 14 11:10:10 crc kubenswrapper[4904]: E0214 11:10:10.378363 4904 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="800ms" Feb 14 11:10:10 crc kubenswrapper[4904]: W0214 11:10:10.570731 4904 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Feb 14 11:10:10 crc kubenswrapper[4904]: E0214 11:10:10.570797 4904 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.609338 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.610221 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.610255 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.610264 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.610341 4904 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 14 11:10:10 crc kubenswrapper[4904]: E0214 11:10:10.610648 4904 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.17:6443: connect: connection refused" node="crc" Feb 14 11:10:10 crc kubenswrapper[4904]: W0214 11:10:10.712110 4904 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Feb 14 11:10:10 crc kubenswrapper[4904]: E0214 11:10:10.712186 4904 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.771396 4904 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.775436 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 22:34:05.37016404 +0000 UTC Feb 14 11:10:10 crc kubenswrapper[4904]: W0214 11:10:10.833829 4904 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Feb 14 11:10:10 crc kubenswrapper[4904]: E0214 11:10:10.834219 4904 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.843537 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9edaf241e93f31ded2eba6b7e6c78b76da34b442735a5c27ec5f7f9fcfb99197"} Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.844373 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6cf3628a9172eed0ac602e5f77596d7a12a3ce7f2c0857fb13ceb66245060f44"} Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.846490 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2a671812bf72b40078117872711b366d6483b90d200a193539c9a1ef87be6c48"} Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.847593 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"14dcae69a98f28901aae49b09877dce7a526f99d8c6229138b0d3672e0d318a6"} Feb 14 11:10:10 crc kubenswrapper[4904]: I0214 11:10:10.849215 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"43fc2fe725b883e9e06518b3a9198c3198fa7f5a0a5ccf8136229d12c1210450"} Feb 14 11:10:10 crc kubenswrapper[4904]: W0214 11:10:10.971760 4904 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Feb 14 11:10:10 crc kubenswrapper[4904]: E0214 11:10:10.971828 4904 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Feb 14 11:10:11 crc kubenswrapper[4904]: E0214 11:10:11.179558 4904 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="1.6s" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.411313 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.412859 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.412891 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.412899 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.412924 4904 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 14 11:10:11 crc kubenswrapper[4904]: E0214 11:10:11.413444 4904 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.17:6443: connect: connection refused" node="crc" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.764948 4904 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 14 11:10:11 crc kubenswrapper[4904]: E0214 11:10:11.765737 4904 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.771204 4904 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.775523 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 18:11:07.315688556 +0000 UTC Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.852881 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f"} Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.853162 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b"} Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.853286 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7"} Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.853369 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2"} Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.852947 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.854261 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.854338 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.854392 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.854418 4904 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ac46ed3d612322b4a0152f7f991ca409737e538ed88bc42da8702d90f814a3a0" exitCode=0 Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.854517 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ac46ed3d612322b4a0152f7f991ca409737e538ed88bc42da8702d90f814a3a0"} Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.854872 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.855753 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.855849 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.855926 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.856790 4904 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="3cabed3e8c25fb0a1231609de9a3c24ff9cff7728385e57a5aa2a895b8288889" exitCode=0 Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.856911 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"3cabed3e8c25fb0a1231609de9a3c24ff9cff7728385e57a5aa2a895b8288889"} Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.857022 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.857901 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.857938 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.857950 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.859496 4904 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5" exitCode=0 Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.859676 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.859682 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5"} Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.860650 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.860717 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.860770 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.868199 4904 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0" exitCode=0 Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.868237 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0"} Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.868337 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.869356 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.869381 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.869391 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.871473 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.887721 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.887774 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:11 crc kubenswrapper[4904]: I0214 11:10:11.887789 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:12 crc kubenswrapper[4904]: W0214 11:10:12.430379 4904 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Feb 14 11:10:12 crc kubenswrapper[4904]: E0214 11:10:12.430442 4904 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Feb 14 11:10:12 crc kubenswrapper[4904]: W0214 11:10:12.474920 4904 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Feb 14 11:10:12 crc kubenswrapper[4904]: E0214 11:10:12.474991 4904 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.771758 4904 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.775615 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 03:37:47.571494379 +0000 UTC Feb 14 11:10:12 crc kubenswrapper[4904]: E0214 11:10:12.780783 4904 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="3.2s" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.872039 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"15013a04b2c95dd2b7d0484a604cb50135f57e166b332e6ce6195fe0a0cdb429"} Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.872157 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.873379 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.873410 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.873422 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.876049 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f2411332719f59b693409f1e2c9a10c280a102c11216e67d8eea8bfa8bdfe809"} Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.876098 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"cc073ab93f3609414e270a023f8d2bad18f489eb6b37ac9a15cd6bf8b4dfbbf4"} Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.876115 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ee16aacae98d845ea68a860c26436a43d376f74ea7abd4396984fa63a836cb49"} Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.876130 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.876982 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.877015 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.877033 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.878578 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75"} Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.878616 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90"} Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.878624 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.878634 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37"} Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.878646 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e"} Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.878658 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634"} Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.879326 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.879349 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.879363 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.880512 4904 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="dee509421e60e93c58c3179fe65cee84f394883ae21fba6b9b659d0bb0ee04aa" exitCode=0 Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.880574 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"dee509421e60e93c58c3179fe65cee84f394883ae21fba6b9b659d0bb0ee04aa"} Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.880610 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.880617 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.881516 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.881537 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.881549 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.881562 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.881580 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:12 crc kubenswrapper[4904]: I0214 11:10:12.881593 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.013788 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.014853 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.014892 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.014903 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.014928 4904 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 14 11:10:13 crc kubenswrapper[4904]: E0214 11:10:13.015336 4904 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.17:6443: connect: connection refused" node="crc" Feb 14 11:10:13 crc kubenswrapper[4904]: E0214 11:10:13.136248 4904 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.17:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1894186cf8a545d7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-14 11:10:09.770563031 +0000 UTC m=+0.583327712,LastTimestamp:2026-02-14 11:10:09.770563031 +0000 UTC m=+0.583327712,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.168426 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.700793 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.775776 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 05:18:49.694773817 +0000 UTC Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.885657 4904 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="8615b2f30fbd8ebc2e6b82db025a7de6c3d282e65ce2505d20662e8cd00931a2" exitCode=0 Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.885919 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.885950 4904 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.885980 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"8615b2f30fbd8ebc2e6b82db025a7de6c3d282e65ce2505d20662e8cd00931a2"} Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.886006 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.886077 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.886472 4904 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.886654 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.887990 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.889139 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.889340 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.889506 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.889450 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.889725 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.889749 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.889510 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.889552 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.890111 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.890140 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.890141 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.890171 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.889391 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.890530 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:13 crc kubenswrapper[4904]: I0214 11:10:13.890558 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:14 crc kubenswrapper[4904]: I0214 11:10:14.776158 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 19:14:06.218942212 +0000 UTC Feb 14 11:10:14 crc kubenswrapper[4904]: I0214 11:10:14.894177 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:14 crc kubenswrapper[4904]: I0214 11:10:14.894178 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4af27db30a138f5ce1be1e3cd8eeaca6ee3a09f059b811802fd7a5368cc9795b"} Feb 14 11:10:14 crc kubenswrapper[4904]: I0214 11:10:14.894224 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5e3a11b5b8c5145431de1e1cb80e5464405a400ee27ec3cb088fa260df0ef2e2"} Feb 14 11:10:14 crc kubenswrapper[4904]: I0214 11:10:14.894243 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6eb3e2d65c5ab826fefb31d048a9070d09e34ed56a8496d8f6f62a5738053934"} Feb 14 11:10:14 crc kubenswrapper[4904]: I0214 11:10:14.894259 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fef0797f82eb4ab705467907eb0c8e315ca990480399be7518749e27ab8c7124"} Feb 14 11:10:14 crc kubenswrapper[4904]: I0214 11:10:14.894277 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d1436ce493ff495cc047dbe763e209e1224f29ce8b77689ea1d44e279dd5c3c5"} Feb 14 11:10:14 crc kubenswrapper[4904]: I0214 11:10:14.894337 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:14 crc kubenswrapper[4904]: I0214 11:10:14.895521 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:14 crc kubenswrapper[4904]: I0214 11:10:14.895541 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:14 crc kubenswrapper[4904]: I0214 11:10:14.895580 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:14 crc kubenswrapper[4904]: I0214 11:10:14.895595 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:14 crc kubenswrapper[4904]: I0214 11:10:14.895547 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:14 crc kubenswrapper[4904]: I0214 11:10:14.895716 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.287460 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.776701 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 16:10:53.048058406 +0000 UTC Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.805189 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.805332 4904 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.805370 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.806404 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.806456 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.806471 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.896224 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.896241 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.897559 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.897731 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.897896 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.897903 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.898047 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.898061 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.947935 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.948025 4904 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.948051 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.948903 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.948955 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.948975 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:15 crc kubenswrapper[4904]: I0214 11:10:15.999479 4904 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 14 11:10:16 crc kubenswrapper[4904]: I0214 11:10:16.083063 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 14 11:10:16 crc kubenswrapper[4904]: I0214 11:10:16.215864 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:16 crc kubenswrapper[4904]: I0214 11:10:16.220645 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:16 crc kubenswrapper[4904]: I0214 11:10:16.220719 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:16 crc kubenswrapper[4904]: I0214 11:10:16.220735 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:16 crc kubenswrapper[4904]: I0214 11:10:16.220910 4904 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 14 11:10:16 crc kubenswrapper[4904]: I0214 11:10:16.701906 4904 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 14 11:10:16 crc kubenswrapper[4904]: I0214 11:10:16.702036 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 14 11:10:16 crc kubenswrapper[4904]: I0214 11:10:16.776942 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 04:47:20.797864522 +0000 UTC Feb 14 11:10:16 crc kubenswrapper[4904]: I0214 11:10:16.892594 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 14 11:10:16 crc kubenswrapper[4904]: I0214 11:10:16.898786 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:16 crc kubenswrapper[4904]: I0214 11:10:16.900522 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:16 crc kubenswrapper[4904]: I0214 11:10:16.900627 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:16 crc kubenswrapper[4904]: I0214 11:10:16.900653 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:17 crc kubenswrapper[4904]: I0214 11:10:17.312575 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 14 11:10:17 crc kubenswrapper[4904]: I0214 11:10:17.312930 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:17 crc kubenswrapper[4904]: I0214 11:10:17.314417 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:17 crc kubenswrapper[4904]: I0214 11:10:17.314488 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:17 crc kubenswrapper[4904]: I0214 11:10:17.314518 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:17 crc kubenswrapper[4904]: I0214 11:10:17.778026 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 07:03:19.78503417 +0000 UTC Feb 14 11:10:17 crc kubenswrapper[4904]: I0214 11:10:17.902028 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:17 crc kubenswrapper[4904]: I0214 11:10:17.903518 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:17 crc kubenswrapper[4904]: I0214 11:10:17.903574 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:17 crc kubenswrapper[4904]: I0214 11:10:17.903597 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:18 crc kubenswrapper[4904]: I0214 11:10:18.779011 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 11:13:26.161482227 +0000 UTC Feb 14 11:10:18 crc kubenswrapper[4904]: I0214 11:10:18.814386 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:10:18 crc kubenswrapper[4904]: I0214 11:10:18.814923 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:18 crc kubenswrapper[4904]: I0214 11:10:18.816483 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:18 crc kubenswrapper[4904]: I0214 11:10:18.816522 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:18 crc kubenswrapper[4904]: I0214 11:10:18.816532 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:19 crc kubenswrapper[4904]: I0214 11:10:19.089158 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:10:19 crc kubenswrapper[4904]: I0214 11:10:19.089341 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:19 crc kubenswrapper[4904]: I0214 11:10:19.090186 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:19 crc kubenswrapper[4904]: I0214 11:10:19.090207 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:19 crc kubenswrapper[4904]: I0214 11:10:19.090216 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:19 crc kubenswrapper[4904]: I0214 11:10:19.095420 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:10:19 crc kubenswrapper[4904]: I0214 11:10:19.779377 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 17:10:23.491309333 +0000 UTC Feb 14 11:10:19 crc kubenswrapper[4904]: I0214 11:10:19.908188 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:19 crc kubenswrapper[4904]: I0214 11:10:19.910220 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:19 crc kubenswrapper[4904]: I0214 11:10:19.910314 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:19 crc kubenswrapper[4904]: I0214 11:10:19.910385 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:19 crc kubenswrapper[4904]: E0214 11:10:19.914398 4904 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 14 11:10:20 crc kubenswrapper[4904]: I0214 11:10:20.779971 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 18:17:38.338042172 +0000 UTC Feb 14 11:10:21 crc kubenswrapper[4904]: I0214 11:10:21.781119 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 07:56:19.668626238 +0000 UTC Feb 14 11:10:22 crc kubenswrapper[4904]: I0214 11:10:22.781652 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 16:11:23.022666414 +0000 UTC Feb 14 11:10:23 crc kubenswrapper[4904]: W0214 11:10:23.529008 4904 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 14 11:10:23 crc kubenswrapper[4904]: I0214 11:10:23.529196 4904 trace.go:236] Trace[1587031926]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Feb-2026 11:10:13.527) (total time: 10001ms): Feb 14 11:10:23 crc kubenswrapper[4904]: Trace[1587031926]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (11:10:23.528) Feb 14 11:10:23 crc kubenswrapper[4904]: Trace[1587031926]: [10.001984537s] [10.001984537s] END Feb 14 11:10:23 crc kubenswrapper[4904]: E0214 11:10:23.529244 4904 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 14 11:10:23 crc kubenswrapper[4904]: W0214 11:10:23.755968 4904 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 14 11:10:23 crc kubenswrapper[4904]: I0214 11:10:23.756193 4904 trace.go:236] Trace[1653923268]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Feb-2026 11:10:13.754) (total time: 10001ms): Feb 14 11:10:23 crc kubenswrapper[4904]: Trace[1653923268]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (11:10:23.755) Feb 14 11:10:23 crc kubenswrapper[4904]: Trace[1653923268]: [10.001878755s] [10.001878755s] END Feb 14 11:10:23 crc kubenswrapper[4904]: E0214 11:10:23.756249 4904 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 14 11:10:23 crc kubenswrapper[4904]: I0214 11:10:23.772367 4904 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 14 11:10:23 crc kubenswrapper[4904]: I0214 11:10:23.782151 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 17:24:56.391144116 +0000 UTC Feb 14 11:10:23 crc kubenswrapper[4904]: I0214 11:10:23.900360 4904 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 14 11:10:23 crc kubenswrapper[4904]: I0214 11:10:23.900438 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 14 11:10:23 crc kubenswrapper[4904]: I0214 11:10:23.936414 4904 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 14 11:10:23 crc kubenswrapper[4904]: I0214 11:10:23.936534 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 14 11:10:24 crc kubenswrapper[4904]: I0214 11:10:24.783159 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 15:58:08.811647104 +0000 UTC Feb 14 11:10:25 crc kubenswrapper[4904]: I0214 11:10:25.294739 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:10:25 crc kubenswrapper[4904]: I0214 11:10:25.294934 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:25 crc kubenswrapper[4904]: I0214 11:10:25.296320 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:25 crc kubenswrapper[4904]: I0214 11:10:25.296354 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:25 crc kubenswrapper[4904]: I0214 11:10:25.296363 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:25 crc kubenswrapper[4904]: I0214 11:10:25.783960 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 03:01:11.860661665 +0000 UTC Feb 14 11:10:25 crc kubenswrapper[4904]: I0214 11:10:25.811426 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:10:25 crc kubenswrapper[4904]: I0214 11:10:25.811691 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:25 crc kubenswrapper[4904]: I0214 11:10:25.813265 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:25 crc kubenswrapper[4904]: I0214 11:10:25.813307 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:25 crc kubenswrapper[4904]: I0214 11:10:25.813321 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:25 crc kubenswrapper[4904]: I0214 11:10:25.816553 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:10:25 crc kubenswrapper[4904]: I0214 11:10:25.923571 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:25 crc kubenswrapper[4904]: I0214 11:10:25.924487 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:25 crc kubenswrapper[4904]: I0214 11:10:25.924521 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:25 crc kubenswrapper[4904]: I0214 11:10:25.924530 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:26 crc kubenswrapper[4904]: I0214 11:10:26.112412 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 14 11:10:26 crc kubenswrapper[4904]: I0214 11:10:26.112607 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:26 crc kubenswrapper[4904]: I0214 11:10:26.113823 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:26 crc kubenswrapper[4904]: I0214 11:10:26.113868 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:26 crc kubenswrapper[4904]: I0214 11:10:26.113877 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:26 crc kubenswrapper[4904]: I0214 11:10:26.124695 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 14 11:10:26 crc kubenswrapper[4904]: I0214 11:10:26.701677 4904 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 14 11:10:26 crc kubenswrapper[4904]: I0214 11:10:26.702313 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 14 11:10:26 crc kubenswrapper[4904]: I0214 11:10:26.784089 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 17:03:35.95511283 +0000 UTC Feb 14 11:10:26 crc kubenswrapper[4904]: I0214 11:10:26.926143 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:26 crc kubenswrapper[4904]: I0214 11:10:26.927124 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:26 crc kubenswrapper[4904]: I0214 11:10:26.927157 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:26 crc kubenswrapper[4904]: I0214 11:10:26.927165 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:27 crc kubenswrapper[4904]: I0214 11:10:27.785135 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 07:22:46.634947909 +0000 UTC Feb 14 11:10:28 crc kubenswrapper[4904]: I0214 11:10:28.321462 4904 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 14 11:10:28 crc kubenswrapper[4904]: I0214 11:10:28.785530 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 05:51:04.654265378 +0000 UTC Feb 14 11:10:28 crc kubenswrapper[4904]: I0214 11:10:28.914422 4904 trace.go:236] Trace[1933501358]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Feb-2026 11:10:16.166) (total time: 12748ms): Feb 14 11:10:28 crc kubenswrapper[4904]: Trace[1933501358]: ---"Objects listed" error: 12748ms (11:10:28.914) Feb 14 11:10:28 crc kubenswrapper[4904]: Trace[1933501358]: [12.748100971s] [12.748100971s] END Feb 14 11:10:28 crc kubenswrapper[4904]: I0214 11:10:28.914453 4904 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 14 11:10:28 crc kubenswrapper[4904]: E0214 11:10:28.915935 4904 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Feb 14 11:10:28 crc kubenswrapper[4904]: I0214 11:10:28.919922 4904 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 14 11:10:28 crc kubenswrapper[4904]: I0214 11:10:28.921521 4904 trace.go:236] Trace[870549889]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Feb-2026 11:10:17.785) (total time: 11136ms): Feb 14 11:10:28 crc kubenswrapper[4904]: Trace[870549889]: ---"Objects listed" error: 11136ms (11:10:28.921) Feb 14 11:10:28 crc kubenswrapper[4904]: Trace[870549889]: [11.136189652s] [11.136189652s] END Feb 14 11:10:28 crc kubenswrapper[4904]: I0214 11:10:28.921667 4904 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 14 11:10:28 crc kubenswrapper[4904]: E0214 11:10:28.925121 4904 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Feb 14 11:10:28 crc kubenswrapper[4904]: I0214 11:10:28.927998 4904 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 14 11:10:28 crc kubenswrapper[4904]: I0214 11:10:28.943761 4904 csr.go:261] certificate signing request csr-f2fmh is approved, waiting to be issued Feb 14 11:10:28 crc kubenswrapper[4904]: I0214 11:10:28.952050 4904 csr.go:257] certificate signing request csr-f2fmh is issued Feb 14 11:10:28 crc kubenswrapper[4904]: I0214 11:10:28.955760 4904 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:32980->192.168.126.11:17697: read: connection reset by peer" start-of-body= Feb 14 11:10:28 crc kubenswrapper[4904]: I0214 11:10:28.956078 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:32980->192.168.126.11:17697: read: connection reset by peer" Feb 14 11:10:28 crc kubenswrapper[4904]: I0214 11:10:28.955973 4904 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:32988->192.168.126.11:17697: read: connection reset by peer" start-of-body= Feb 14 11:10:28 crc kubenswrapper[4904]: I0214 11:10:28.956654 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:32988->192.168.126.11:17697: read: connection reset by peer" Feb 14 11:10:28 crc kubenswrapper[4904]: I0214 11:10:28.957120 4904 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 14 11:10:28 crc kubenswrapper[4904]: I0214 11:10:28.957215 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 14 11:10:28 crc kubenswrapper[4904]: I0214 11:10:28.957642 4904 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 14 11:10:28 crc kubenswrapper[4904]: I0214 11:10:28.957667 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.200019 4904 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.672220 4904 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 14 11:10:29 crc kubenswrapper[4904]: E0214 11:10:29.672488 4904 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/events\": read tcp 38.102.83.17:42734->38.102.83.17:6443: use of closed network connection" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.1894186d19a3bdc1 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-14 11:10:10.324110785 +0000 UTC m=+1.136875446,LastTimestamp:2026-02-14 11:10:10.324110785 +0000 UTC m=+1.136875446,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 14 11:10:29 crc kubenswrapper[4904]: W0214 11:10:29.672597 4904 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.CSIDriver ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Feb 14 11:10:29 crc kubenswrapper[4904]: W0214 11:10:29.672627 4904 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.Service ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Feb 14 11:10:29 crc kubenswrapper[4904]: W0214 11:10:29.672734 4904 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.Node ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.772200 4904 apiserver.go:52] "Watching apiserver" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.776745 4904 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.777271 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.777796 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.777957 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:29 crc kubenswrapper[4904]: E0214 11:10:29.778027 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.778203 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:29 crc kubenswrapper[4904]: E0214 11:10:29.778285 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.778341 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.778410 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.778563 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:29 crc kubenswrapper[4904]: E0214 11:10:29.778653 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.781959 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.781980 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.781990 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.781961 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.782074 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.781960 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.782492 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.782514 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.785862 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 00:06:29.339011991 +0000 UTC Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.786801 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.819420 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.831663 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.859405 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.872070 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.876975 4904 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.888390 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.902792 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926169 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926214 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926238 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926259 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926279 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926298 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926320 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926341 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926360 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926378 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926398 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926418 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926438 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926457 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926472 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926486 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926499 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926528 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926550 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926573 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926590 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926605 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926619 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926643 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926664 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926684 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926698 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926717 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926737 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926759 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926778 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926800 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926821 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926858 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926886 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926923 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926955 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926976 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.926994 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927033 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927055 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927075 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927098 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927120 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927141 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927176 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927201 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927223 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927256 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927292 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927314 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927336 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927356 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927379 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927401 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927421 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927441 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927463 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927486 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927509 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927531 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927552 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927572 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927593 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927613 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927634 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927654 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927675 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927700 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927721 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927741 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927764 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927785 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927808 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927848 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927874 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927897 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927918 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927939 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927960 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.927980 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928003 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928026 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928043 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928059 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928075 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928090 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928108 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928123 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928141 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928164 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928183 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928199 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928215 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928237 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928258 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928280 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928300 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928320 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928343 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928363 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928384 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928404 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928423 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928443 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928463 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928483 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928506 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928527 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928548 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928569 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928588 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928611 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928634 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928655 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928677 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928698 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928720 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928742 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928765 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928785 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928805 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928824 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928864 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928887 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928908 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928931 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928953 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928973 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.928994 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929017 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929039 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929061 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929084 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929104 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929125 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929153 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929176 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929199 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929224 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929249 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929275 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929298 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929322 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929343 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929364 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929383 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929406 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929430 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929454 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929478 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929499 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929521 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929619 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929646 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929670 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929691 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929714 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929739 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929763 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929787 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929808 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929848 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929873 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929896 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929920 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929944 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929966 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.929990 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930011 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930036 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930060 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930100 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930126 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930152 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930178 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930202 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930225 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930251 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930273 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930297 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930323 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930348 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930372 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930400 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930425 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930448 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930474 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930499 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930522 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930545 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930568 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930592 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930616 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930639 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930687 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930719 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930746 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930772 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930798 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930823 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930868 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930895 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930922 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930954 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.930979 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.931006 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.931028 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.931046 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.931895 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.932054 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.932090 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.932067 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.932262 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.932389 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.932412 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.932431 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.932653 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.932688 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.932740 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.932855 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.932993 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.933024 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.933056 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.933206 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.933347 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.933430 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.933457 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.933581 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.933675 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.933824 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.934053 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.934237 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.934452 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.934752 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.935197 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.935385 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.935893 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.936078 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.936210 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.936370 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.936552 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.936758 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.936822 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.936926 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.937017 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.937325 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.937618 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.938390 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.938396 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.938419 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.938627 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.939051 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.939216 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.939458 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.939494 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.939675 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.939688 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.939894 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.939919 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.940827 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.941354 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.941419 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.941904 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.941996 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.942165 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.942562 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.942753 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.942956 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.943352 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: E0214 11:10:29.943435 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:10:30.443417676 +0000 UTC m=+21.256182337 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.943604 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.943772 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.944392 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.944587 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.944720 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.944942 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.949995 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.950033 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.950123 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.950283 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.950310 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.951121 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.951578 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.952343 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.952947 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.953050 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.953245 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.953277 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.953412 4904 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-14 11:05:28 +0000 UTC, rotation deadline is 2026-12-02 21:17:21.622432141 +0000 UTC Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.953447 4904 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6994h6m51.668987859s for next certificate rotation Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.953452 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.953491 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.953461 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.954534 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.954976 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.955025 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.955381 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.955516 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.956042 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.956180 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.956415 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.956476 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.956337 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.957540 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.956774 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.957102 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.957313 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.957767 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.958085 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.958376 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.958463 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.958765 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.959024 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.959332 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: E0214 11:10:29.959471 4904 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.959543 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: E0214 11:10:29.959574 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:30.459549566 +0000 UTC m=+21.272314227 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 14 11:10:29 crc kubenswrapper[4904]: E0214 11:10:29.959763 4904 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 14 11:10:29 crc kubenswrapper[4904]: E0214 11:10:29.959803 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:30.459794941 +0000 UTC m=+21.272559602 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.960809 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.961024 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.961084 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.961349 4904 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.961572 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.961299 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.961284 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.961326 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.961371 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.961772 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.961592 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.962239 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.962370 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.965187 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.965584 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.967451 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.965635 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.961387 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.969960 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.970179 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.970199 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.971026 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.971085 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.971232 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.971409 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.971798 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.972400 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.972885 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.964746 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.973329 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.973576 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.973803 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.973881 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.974297 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.974957 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.975176 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.975238 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.975505 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.975804 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.975895 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.976170 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.976232 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.976419 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.976472 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.976809 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.976905 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.977485 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.977792 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.978745 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.979365 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.980125 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.980602 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.981494 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.981607 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.981773 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.981778 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.982051 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.982234 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.982302 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.982528 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.982535 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.982661 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.982907 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.983083 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.983124 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.983181 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.983467 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.985059 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.985255 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: E0214 11:10:29.987519 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 14 11:10:29 crc kubenswrapper[4904]: E0214 11:10:29.987635 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.987684 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: E0214 11:10:29.987714 4904 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:29 crc kubenswrapper[4904]: E0214 11:10:29.987881 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:30.487855481 +0000 UTC m=+21.300620142 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.988072 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.988502 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.988620 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.988808 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.988943 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.988976 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: E0214 11:10:29.989283 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 14 11:10:29 crc kubenswrapper[4904]: E0214 11:10:29.989376 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 14 11:10:29 crc kubenswrapper[4904]: E0214 11:10:29.989447 4904 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:29 crc kubenswrapper[4904]: E0214 11:10:29.989591 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:30.489562361 +0000 UTC m=+21.302327252 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.990464 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.990854 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.992590 4904 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75" exitCode=255 Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.992627 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75"} Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.993151 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.994494 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.994870 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.994924 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.995116 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.995860 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.995901 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.996041 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.996508 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.996960 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.997149 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.997366 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.997680 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.998052 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.998473 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:10:29 crc kubenswrapper[4904]: I0214 11:10:29.998658 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.000036 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.000195 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.001268 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.002093 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.005494 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.006407 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.013541 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.025705 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.026664 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.030602 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.048912 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.049727 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.049844 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.053492 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.053844 4904 scope.go:117] "RemoveContainer" containerID="e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054118 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054266 4904 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054283 4904 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054293 4904 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054303 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054312 4904 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054320 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054329 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054337 4904 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054346 4904 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054354 4904 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054363 4904 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054371 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054380 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054389 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054566 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054659 4904 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054670 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054679 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054688 4904 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054696 4904 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054704 4904 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054713 4904 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054723 4904 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054731 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054739 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054755 4904 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054764 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054773 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054781 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054790 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054798 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054808 4904 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054820 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054849 4904 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054862 4904 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054897 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054917 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054931 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054943 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054955 4904 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054966 4904 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054977 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054988 4904 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.054999 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055009 4904 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055030 4904 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055041 4904 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055052 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055062 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055073 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055085 4904 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055098 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055109 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055119 4904 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055127 4904 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055136 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055143 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055153 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055161 4904 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055178 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055187 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055195 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055203 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055212 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055221 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055229 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055238 4904 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055246 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055254 4904 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055262 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055273 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055299 4904 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055308 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055315 4904 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055324 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055335 4904 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055343 4904 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055351 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055359 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055369 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055378 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055386 4904 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055395 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055405 4904 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055413 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055422 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055430 4904 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055438 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055446 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055454 4904 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055463 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055475 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055483 4904 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055491 4904 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055499 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055508 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055516 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055525 4904 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055534 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055542 4904 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055551 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055559 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055567 4904 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055575 4904 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055583 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055592 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055600 4904 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055608 4904 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055617 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055626 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055634 4904 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055642 4904 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055650 4904 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055658 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055667 4904 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055691 4904 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055700 4904 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055708 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055717 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055725 4904 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055733 4904 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055740 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055748 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055759 4904 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055767 4904 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055775 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055782 4904 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055790 4904 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055799 4904 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055807 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055814 4904 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055822 4904 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055846 4904 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055854 4904 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055862 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055871 4904 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055879 4904 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055887 4904 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055895 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055903 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055911 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055919 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055927 4904 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055935 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055942 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055950 4904 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055958 4904 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055967 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055976 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055986 4904 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.055994 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056001 4904 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056009 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056017 4904 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056027 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056037 4904 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056045 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056053 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056062 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056070 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056078 4904 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056087 4904 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056095 4904 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056102 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056111 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056119 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056128 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056136 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056146 4904 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056155 4904 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056163 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056171 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056180 4904 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056187 4904 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056196 4904 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056203 4904 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056212 4904 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056220 4904 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056230 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056239 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056248 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056256 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056264 4904 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056271 4904 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056279 4904 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056287 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056295 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056305 4904 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056314 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056321 4904 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056330 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056339 4904 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056347 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.056355 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.076092 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.090947 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.099718 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.107677 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.118080 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.146096 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.160805 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.176102 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.192762 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.207468 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.228256 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.243080 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.267245 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.293056 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.459345 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:10:30 crc kubenswrapper[4904]: E0214 11:10:30.459561 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:10:31.459544071 +0000 UTC m=+22.272308732 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.559963 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:30 crc kubenswrapper[4904]: E0214 11:10:30.560108 4904 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 14 11:10:30 crc kubenswrapper[4904]: E0214 11:10:30.560308 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:31.56028765 +0000 UTC m=+22.373052371 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.560232 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.560378 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.560428 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:30 crc kubenswrapper[4904]: E0214 11:10:30.560512 4904 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 14 11:10:30 crc kubenswrapper[4904]: E0214 11:10:30.560525 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 14 11:10:30 crc kubenswrapper[4904]: E0214 11:10:30.560541 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 14 11:10:30 crc kubenswrapper[4904]: E0214 11:10:30.560551 4904 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:30 crc kubenswrapper[4904]: E0214 11:10:30.560558 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:31.560547636 +0000 UTC m=+22.373312297 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 14 11:10:30 crc kubenswrapper[4904]: E0214 11:10:30.560578 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:31.560570016 +0000 UTC m=+22.373334677 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:30 crc kubenswrapper[4904]: E0214 11:10:30.560799 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 14 11:10:30 crc kubenswrapper[4904]: E0214 11:10:30.560910 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 14 11:10:30 crc kubenswrapper[4904]: E0214 11:10:30.560993 4904 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:30 crc kubenswrapper[4904]: E0214 11:10:30.561109 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:31.561095628 +0000 UTC m=+22.373860289 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.564014 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-zjhq2"] Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.564446 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-zjhq2" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.566221 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.566885 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.567036 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.585046 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.609171 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.624127 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.645051 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.661694 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/95999cbb-5c5a-4dd9-afbd-6472f52bf66b-hosts-file\") pod \"node-resolver-zjhq2\" (UID: \"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\") " pod="openshift-dns/node-resolver-zjhq2" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.661979 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv4sg\" (UniqueName: \"kubernetes.io/projected/95999cbb-5c5a-4dd9-afbd-6472f52bf66b-kube-api-access-vv4sg\") pod \"node-resolver-zjhq2\" (UID: \"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\") " pod="openshift-dns/node-resolver-zjhq2" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.670536 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.697186 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.731476 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.762473 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/95999cbb-5c5a-4dd9-afbd-6472f52bf66b-hosts-file\") pod \"node-resolver-zjhq2\" (UID: \"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\") " pod="openshift-dns/node-resolver-zjhq2" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.762509 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv4sg\" (UniqueName: \"kubernetes.io/projected/95999cbb-5c5a-4dd9-afbd-6472f52bf66b-kube-api-access-vv4sg\") pod \"node-resolver-zjhq2\" (UID: \"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\") " pod="openshift-dns/node-resolver-zjhq2" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.762771 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/95999cbb-5c5a-4dd9-afbd-6472f52bf66b-hosts-file\") pod \"node-resolver-zjhq2\" (UID: \"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\") " pod="openshift-dns/node-resolver-zjhq2" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.767605 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.786432 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 07:27:58.037310897 +0000 UTC Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.794452 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv4sg\" (UniqueName: \"kubernetes.io/projected/95999cbb-5c5a-4dd9-afbd-6472f52bf66b-kube-api-access-vv4sg\") pod \"node-resolver-zjhq2\" (UID: \"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\") " pod="openshift-dns/node-resolver-zjhq2" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.876906 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-zjhq2" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.933329 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-tnf2t"] Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.933648 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.941850 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.941970 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.941981 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.953209 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.953399 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 14 11:10:30 crc kubenswrapper[4904]: I0214 11:10:30.977701 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.009270 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"5b1612b6ecf821c94a2f6158c1a38d7245854261be9f9f4fe7c92ac708844d82"} Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.010648 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f"} Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.010733 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"79290019fe3ad8bf8e3f37810452d8df2f6142a6a00214b0c62d35b4a7348cae"} Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.013578 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-zjhq2" event={"ID":"95999cbb-5c5a-4dd9-afbd-6472f52bf66b","Type":"ContainerStarted","Data":"75ec930f0801a362b9e1552d118441985ab6dffd3080afd8310fb0b364744123"} Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.015675 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.023209 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b"} Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.023252 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4"} Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.023261 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"87833fea67dae75caeaf3e19605387488068c6e7115dec38202da823ff5834fe"} Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.033343 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.034197 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.044109 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.045962 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f"} Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.046399 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.058574 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.064970 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqg5c\" (UniqueName: \"kubernetes.io/projected/7790d276-2506-4641-8c22-67e6441417f0-kube-api-access-dqg5c\") pod \"machine-config-daemon-tnf2t\" (UID: \"7790d276-2506-4641-8c22-67e6441417f0\") " pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.065038 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7790d276-2506-4641-8c22-67e6441417f0-rootfs\") pod \"machine-config-daemon-tnf2t\" (UID: \"7790d276-2506-4641-8c22-67e6441417f0\") " pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.065063 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7790d276-2506-4641-8c22-67e6441417f0-proxy-tls\") pod \"machine-config-daemon-tnf2t\" (UID: \"7790d276-2506-4641-8c22-67e6441417f0\") " pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.065135 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7790d276-2506-4641-8c22-67e6441417f0-mcd-auth-proxy-config\") pod \"machine-config-daemon-tnf2t\" (UID: \"7790d276-2506-4641-8c22-67e6441417f0\") " pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.073580 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.088888 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.097428 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.106141 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.119860 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.133153 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.148667 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.163659 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.165513 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqg5c\" (UniqueName: \"kubernetes.io/projected/7790d276-2506-4641-8c22-67e6441417f0-kube-api-access-dqg5c\") pod \"machine-config-daemon-tnf2t\" (UID: \"7790d276-2506-4641-8c22-67e6441417f0\") " pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.165568 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7790d276-2506-4641-8c22-67e6441417f0-rootfs\") pod \"machine-config-daemon-tnf2t\" (UID: \"7790d276-2506-4641-8c22-67e6441417f0\") " pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.165616 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7790d276-2506-4641-8c22-67e6441417f0-proxy-tls\") pod \"machine-config-daemon-tnf2t\" (UID: \"7790d276-2506-4641-8c22-67e6441417f0\") " pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.165688 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7790d276-2506-4641-8c22-67e6441417f0-mcd-auth-proxy-config\") pod \"machine-config-daemon-tnf2t\" (UID: \"7790d276-2506-4641-8c22-67e6441417f0\") " pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.165961 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7790d276-2506-4641-8c22-67e6441417f0-rootfs\") pod \"machine-config-daemon-tnf2t\" (UID: \"7790d276-2506-4641-8c22-67e6441417f0\") " pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.166859 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7790d276-2506-4641-8c22-67e6441417f0-mcd-auth-proxy-config\") pod \"machine-config-daemon-tnf2t\" (UID: \"7790d276-2506-4641-8c22-67e6441417f0\") " pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.174933 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7790d276-2506-4641-8c22-67e6441417f0-proxy-tls\") pod \"machine-config-daemon-tnf2t\" (UID: \"7790d276-2506-4641-8c22-67e6441417f0\") " pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.182493 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.191325 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqg5c\" (UniqueName: \"kubernetes.io/projected/7790d276-2506-4641-8c22-67e6441417f0-kube-api-access-dqg5c\") pod \"machine-config-daemon-tnf2t\" (UID: \"7790d276-2506-4641-8c22-67e6441417f0\") " pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.199363 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.207002 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.219098 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.239443 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.242796 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:10:31 crc kubenswrapper[4904]: W0214 11:10:31.262283 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7790d276_2506_4641_8c22_67e6441417f0.slice/crio-9d905ef6cac2a3340b9899c25a3d8107d3eb2fd0a60d17fa1bb6c098e90f96ef WatchSource:0}: Error finding container 9d905ef6cac2a3340b9899c25a3d8107d3eb2fd0a60d17fa1bb6c098e90f96ef: Status 404 returned error can't find the container with id 9d905ef6cac2a3340b9899c25a3d8107d3eb2fd0a60d17fa1bb6c098e90f96ef Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.320319 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-jz5f5"] Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.320670 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.321620 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-65flh"] Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.322066 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.322581 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.322749 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.323909 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.324091 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.324687 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.324751 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.328055 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2gwlw"] Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.329923 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.336977 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.337406 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.337542 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.337598 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.337816 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.337886 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.338190 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.338270 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.346166 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.369421 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.391085 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.408713 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.425102 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.450444 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.465468 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.467946 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468121 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-kubelet\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: E0214 11:10:31.468147 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:10:33.468115363 +0000 UTC m=+24.280880024 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468250 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-slash\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468308 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-host-run-netns\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468332 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccqlc\" (UniqueName: \"kubernetes.io/projected/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-kube-api-access-ccqlc\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468357 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1ca5acc0-8b15-4052-883f-1987b7ca0067-ovnkube-config\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468393 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-system-cni-dir\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468412 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-multus-socket-dir-parent\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468432 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-host-var-lib-cni-multus\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468451 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-etc-openvswitch\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468531 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468605 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-etc-kubernetes\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468626 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4edce510-9ef8-498a-8b5c-9b9eb41847eb-cnibin\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468647 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-multus-cni-dir\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468666 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-cnibin\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468688 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-cni-binary-copy\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468717 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-multus-daemon-config\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468733 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-run-netns\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468747 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-run-systemd\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468773 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-var-lib-openvswitch\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468794 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rxmn\" (UniqueName: \"kubernetes.io/projected/4edce510-9ef8-498a-8b5c-9b9eb41847eb-kube-api-access-4rxmn\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468813 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-multus-conf-dir\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468870 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1ca5acc0-8b15-4052-883f-1987b7ca0067-env-overrides\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468906 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-log-socket\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468934 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-host-var-lib-kubelet\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468953 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4edce510-9ef8-498a-8b5c-9b9eb41847eb-cni-binary-copy\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468969 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-host-run-k8s-cni-cncf-io\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.468986 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-cni-netd\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.469005 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/4edce510-9ef8-498a-8b5c-9b9eb41847eb-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.469056 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-hostroot\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.469079 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-node-log\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.469114 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-os-release\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.469140 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-host-var-lib-cni-bin\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.469165 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-host-run-multus-certs\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.469181 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-systemd-units\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.469237 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-run-openvswitch\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.469269 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-run-ovn\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.469294 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-run-ovn-kubernetes\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.469312 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-cni-bin\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.469331 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1ca5acc0-8b15-4052-883f-1987b7ca0067-ovn-node-metrics-cert\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.469348 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1ca5acc0-8b15-4052-883f-1987b7ca0067-ovnkube-script-lib\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.469366 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4edce510-9ef8-498a-8b5c-9b9eb41847eb-system-cni-dir\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.469384 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4edce510-9ef8-498a-8b5c-9b9eb41847eb-tuning-conf-dir\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.469411 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rghsz\" (UniqueName: \"kubernetes.io/projected/1ca5acc0-8b15-4052-883f-1987b7ca0067-kube-api-access-rghsz\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.469442 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4edce510-9ef8-498a-8b5c-9b9eb41847eb-os-release\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.494825 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.538731 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.557769 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570106 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-etc-kubernetes\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570145 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4edce510-9ef8-498a-8b5c-9b9eb41847eb-cnibin\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570162 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-multus-cni-dir\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570185 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-cnibin\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570203 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-cni-binary-copy\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570218 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-multus-daemon-config\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570235 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-run-netns\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570248 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-run-systemd\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570246 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-etc-kubernetes\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570266 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-var-lib-openvswitch\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570285 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rxmn\" (UniqueName: \"kubernetes.io/projected/4edce510-9ef8-498a-8b5c-9b9eb41847eb-kube-api-access-4rxmn\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570300 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-multus-conf-dir\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570313 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1ca5acc0-8b15-4052-883f-1987b7ca0067-env-overrides\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570372 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570387 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-log-socket\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570404 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-host-var-lib-kubelet\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570419 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4edce510-9ef8-498a-8b5c-9b9eb41847eb-cni-binary-copy\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570434 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-host-run-k8s-cni-cncf-io\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570449 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-cni-netd\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570480 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/4edce510-9ef8-498a-8b5c-9b9eb41847eb-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570497 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570514 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570546 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-hostroot\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570562 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-node-log\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570576 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-os-release\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570591 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-host-var-lib-cni-bin\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570608 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-host-run-multus-certs\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570624 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-systemd-units\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570641 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-run-openvswitch\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570659 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-run-ovn\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570675 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-run-ovn-kubernetes\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570689 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-cni-bin\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570702 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1ca5acc0-8b15-4052-883f-1987b7ca0067-ovn-node-metrics-cert\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570716 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1ca5acc0-8b15-4052-883f-1987b7ca0067-ovnkube-script-lib\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570730 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4edce510-9ef8-498a-8b5c-9b9eb41847eb-system-cni-dir\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570753 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4edce510-9ef8-498a-8b5c-9b9eb41847eb-tuning-conf-dir\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570772 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rghsz\" (UniqueName: \"kubernetes.io/projected/1ca5acc0-8b15-4052-883f-1987b7ca0067-kube-api-access-rghsz\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570787 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4edce510-9ef8-498a-8b5c-9b9eb41847eb-os-release\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570802 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-kubelet\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570815 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-slash\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570844 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570860 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-host-run-netns\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570891 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccqlc\" (UniqueName: \"kubernetes.io/projected/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-kube-api-access-ccqlc\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570906 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1ca5acc0-8b15-4052-883f-1987b7ca0067-ovnkube-config\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570922 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570938 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-system-cni-dir\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570951 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-multus-socket-dir-parent\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570959 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-cni-binary-copy\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570965 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-host-var-lib-cni-multus\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.570980 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-etc-openvswitch\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.571000 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4edce510-9ef8-498a-8b5c-9b9eb41847eb-cnibin\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.571038 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-etc-openvswitch\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.571134 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-multus-cni-dir\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.571165 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-cnibin\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.571187 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-systemd-units\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.571208 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-run-netns\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.571228 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-run-systemd\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.571248 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-var-lib-openvswitch\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.571498 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-multus-conf-dir\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.571570 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-multus-daemon-config\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.571607 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-run-openvswitch\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.571640 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-run-ovn\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.571677 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-run-ovn-kubernetes\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.571707 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-cni-bin\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.571931 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1ca5acc0-8b15-4052-883f-1987b7ca0067-env-overrides\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: E0214 11:10:31.571994 4904 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 14 11:10:31 crc kubenswrapper[4904]: E0214 11:10:31.572037 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:33.572026126 +0000 UTC m=+24.384790787 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 14 11:10:31 crc kubenswrapper[4904]: E0214 11:10:31.572136 4904 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 14 11:10:31 crc kubenswrapper[4904]: E0214 11:10:31.572184 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:33.57217186 +0000 UTC m=+24.384936521 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572210 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-slash\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572246 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-log-socket\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572244 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4edce510-9ef8-498a-8b5c-9b9eb41847eb-system-cni-dir\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572275 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-host-var-lib-kubelet\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572233 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-kubelet\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572378 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4edce510-9ef8-498a-8b5c-9b9eb41847eb-os-release\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572445 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-hostroot\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572685 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-host-run-k8s-cni-cncf-io\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572712 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-cni-netd\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572742 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1ca5acc0-8b15-4052-883f-1987b7ca0067-ovnkube-script-lib\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572794 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4edce510-9ef8-498a-8b5c-9b9eb41847eb-cni-binary-copy\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: E0214 11:10:31.572811 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 14 11:10:31 crc kubenswrapper[4904]: E0214 11:10:31.572823 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572825 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: E0214 11:10:31.572849 4904 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:31 crc kubenswrapper[4904]: E0214 11:10:31.572879 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 14 11:10:31 crc kubenswrapper[4904]: E0214 11:10:31.572888 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 14 11:10:31 crc kubenswrapper[4904]: E0214 11:10:31.572892 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:33.572882136 +0000 UTC m=+24.385646797 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:31 crc kubenswrapper[4904]: E0214 11:10:31.572897 4904 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572887 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-host-run-netns\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: E0214 11:10:31.572923 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:33.572915157 +0000 UTC m=+24.385679908 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572922 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-system-cni-dir\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572938 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-host-run-multus-certs\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572951 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-os-release\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572959 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-host-var-lib-cni-bin\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572961 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-node-log\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572973 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-host-var-lib-cni-multus\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.572988 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-multus-socket-dir-parent\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.573012 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1ca5acc0-8b15-4052-883f-1987b7ca0067-ovnkube-config\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.573215 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/4edce510-9ef8-498a-8b5c-9b9eb41847eb-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.574161 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4edce510-9ef8-498a-8b5c-9b9eb41847eb-tuning-conf-dir\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.578200 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1ca5acc0-8b15-4052-883f-1987b7ca0067-ovn-node-metrics-cert\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.596274 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rxmn\" (UniqueName: \"kubernetes.io/projected/4edce510-9ef8-498a-8b5c-9b9eb41847eb-kube-api-access-4rxmn\") pod \"multus-additional-cni-plugins-65flh\" (UID: \"4edce510-9ef8-498a-8b5c-9b9eb41847eb\") " pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.603371 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rghsz\" (UniqueName: \"kubernetes.io/projected/1ca5acc0-8b15-4052-883f-1987b7ca0067-kube-api-access-rghsz\") pod \"ovnkube-node-2gwlw\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.604613 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.615381 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccqlc\" (UniqueName: \"kubernetes.io/projected/0b01cc02-2e62-46e1-b07d-b3c0bff2b050-kube-api-access-ccqlc\") pod \"multus-jz5f5\" (UID: \"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\") " pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.633154 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.635778 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-jz5f5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.647008 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-65flh" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.651257 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.671306 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: W0214 11:10:31.675093 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ca5acc0_8b15_4052_883f_1987b7ca0067.slice/crio-410843315a7fa3be626cd1a1f58bd0f7ad7bfa79e8cf2e69e8a8103c5e86f49a WatchSource:0}: Error finding container 410843315a7fa3be626cd1a1f58bd0f7ad7bfa79e8cf2e69e8a8103c5e86f49a: Status 404 returned error can't find the container with id 410843315a7fa3be626cd1a1f58bd0f7ad7bfa79e8cf2e69e8a8103c5e86f49a Feb 14 11:10:31 crc kubenswrapper[4904]: W0214 11:10:31.676075 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4edce510_9ef8_498a_8b5c_9b9eb41847eb.slice/crio-e69918424682d5df7fa505cf93b4ffe05040d854fb8649166be4781f39e27488 WatchSource:0}: Error finding container e69918424682d5df7fa505cf93b4ffe05040d854fb8649166be4781f39e27488: Status 404 returned error can't find the container with id e69918424682d5df7fa505cf93b4ffe05040d854fb8649166be4781f39e27488 Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.695050 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.739094 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.776045 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.786817 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 04:15:37.727855356 +0000 UTC Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.800458 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.819781 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.832565 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.835745 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:31 crc kubenswrapper[4904]: E0214 11:10:31.835881 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.835985 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.836025 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:31 crc kubenswrapper[4904]: E0214 11:10:31.836075 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:10:31 crc kubenswrapper[4904]: E0214 11:10:31.836892 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.844882 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.846894 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.847366 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.848918 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.849697 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.850822 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.851538 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.852272 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.853789 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.854533 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.855690 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.856327 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.857709 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.858424 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.859069 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.860437 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.867161 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.867912 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.868456 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.870067 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.871030 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.871645 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.873020 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.873557 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.875017 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.875551 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.877097 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.877710 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.879452 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.880044 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.880762 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.881415 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.883427 4904 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.883561 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.885606 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.887260 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.887822 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.889759 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.896319 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.897155 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.897171 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:31Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.898501 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.899295 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.900728 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.901474 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.902878 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.903740 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.905384 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.906239 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.907774 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.908679 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.911053 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.912415 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.913677 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.914268 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.914990 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 14 11:10:31 crc kubenswrapper[4904]: I0214 11:10:31.916338 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.051576 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-zjhq2" event={"ID":"95999cbb-5c5a-4dd9-afbd-6472f52bf66b","Type":"ContainerStarted","Data":"89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4"} Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.053207 4904 generic.go:334] "Generic (PLEG): container finished" podID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerID="11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e" exitCode=0 Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.053281 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerDied","Data":"11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e"} Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.053309 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerStarted","Data":"410843315a7fa3be626cd1a1f58bd0f7ad7bfa79e8cf2e69e8a8103c5e86f49a"} Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.054780 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" event={"ID":"4edce510-9ef8-498a-8b5c-9b9eb41847eb","Type":"ContainerStarted","Data":"5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e"} Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.054807 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" event={"ID":"4edce510-9ef8-498a-8b5c-9b9eb41847eb","Type":"ContainerStarted","Data":"e69918424682d5df7fa505cf93b4ffe05040d854fb8649166be4781f39e27488"} Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.056339 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jz5f5" event={"ID":"0b01cc02-2e62-46e1-b07d-b3c0bff2b050","Type":"ContainerStarted","Data":"9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938"} Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.056380 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jz5f5" event={"ID":"0b01cc02-2e62-46e1-b07d-b3c0bff2b050","Type":"ContainerStarted","Data":"46a44d32b4a3710f076b55309c19cb77bb8fb80d443307ada6ff04b7ffb3ff70"} Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.058309 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a"} Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.058335 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228"} Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.058350 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"9d905ef6cac2a3340b9899c25a3d8107d3eb2fd0a60d17fa1bb6c098e90f96ef"} Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.078650 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.098697 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.120544 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.142732 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.161557 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.175487 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.206675 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.219600 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.233162 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.251070 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.267164 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.282958 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.300744 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.322221 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.345965 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.361888 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.388547 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.408365 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.430801 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.446645 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.470437 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.484225 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.498203 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.511601 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.787470 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 05:19:42.446769548 +0000 UTC Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.936000 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-zrwvd"] Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.936562 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zrwvd" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.940255 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.940431 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.940677 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.940460 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.952343 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.969154 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.981785 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:32 crc kubenswrapper[4904]: I0214 11:10:32.995752 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:32Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.007646 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.020018 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.043581 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.063182 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.065800 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerStarted","Data":"047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72"} Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.065854 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerStarted","Data":"d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4"} Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.065865 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerStarted","Data":"a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39"} Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.065874 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerStarted","Data":"b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a"} Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.071161 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda"} Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.073505 4904 generic.go:334] "Generic (PLEG): container finished" podID="4edce510-9ef8-498a-8b5c-9b9eb41847eb" containerID="5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e" exitCode=0 Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.073881 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" event={"ID":"4edce510-9ef8-498a-8b5c-9b9eb41847eb","Type":"ContainerDied","Data":"5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e"} Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.083429 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.092997 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxbls\" (UniqueName: \"kubernetes.io/projected/04ac70df-4f95-4a9e-82d8-164ae3a71623-kube-api-access-pxbls\") pod \"node-ca-zrwvd\" (UID: \"04ac70df-4f95-4a9e-82d8-164ae3a71623\") " pod="openshift-image-registry/node-ca-zrwvd" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.093063 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/04ac70df-4f95-4a9e-82d8-164ae3a71623-serviceca\") pod \"node-ca-zrwvd\" (UID: \"04ac70df-4f95-4a9e-82d8-164ae3a71623\") " pod="openshift-image-registry/node-ca-zrwvd" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.093095 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/04ac70df-4f95-4a9e-82d8-164ae3a71623-host\") pod \"node-ca-zrwvd\" (UID: \"04ac70df-4f95-4a9e-82d8-164ae3a71623\") " pod="openshift-image-registry/node-ca-zrwvd" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.101082 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.121788 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.137890 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.154121 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.177426 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.194492 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxbls\" (UniqueName: \"kubernetes.io/projected/04ac70df-4f95-4a9e-82d8-164ae3a71623-kube-api-access-pxbls\") pod \"node-ca-zrwvd\" (UID: \"04ac70df-4f95-4a9e-82d8-164ae3a71623\") " pod="openshift-image-registry/node-ca-zrwvd" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.194692 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/04ac70df-4f95-4a9e-82d8-164ae3a71623-serviceca\") pod \"node-ca-zrwvd\" (UID: \"04ac70df-4f95-4a9e-82d8-164ae3a71623\") " pod="openshift-image-registry/node-ca-zrwvd" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.195689 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/04ac70df-4f95-4a9e-82d8-164ae3a71623-host\") pod \"node-ca-zrwvd\" (UID: \"04ac70df-4f95-4a9e-82d8-164ae3a71623\") " pod="openshift-image-registry/node-ca-zrwvd" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.195631 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/04ac70df-4f95-4a9e-82d8-164ae3a71623-serviceca\") pod \"node-ca-zrwvd\" (UID: \"04ac70df-4f95-4a9e-82d8-164ae3a71623\") " pod="openshift-image-registry/node-ca-zrwvd" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.196080 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/04ac70df-4f95-4a9e-82d8-164ae3a71623-host\") pod \"node-ca-zrwvd\" (UID: \"04ac70df-4f95-4a9e-82d8-164ae3a71623\") " pod="openshift-image-registry/node-ca-zrwvd" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.199070 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.213818 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.221865 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxbls\" (UniqueName: \"kubernetes.io/projected/04ac70df-4f95-4a9e-82d8-164ae3a71623-kube-api-access-pxbls\") pod \"node-ca-zrwvd\" (UID: \"04ac70df-4f95-4a9e-82d8-164ae3a71623\") " pod="openshift-image-registry/node-ca-zrwvd" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.226743 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.246476 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.268870 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.307424 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.345408 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.385729 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.389011 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zrwvd" Feb 14 11:10:33 crc kubenswrapper[4904]: W0214 11:10:33.405359 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04ac70df_4f95_4a9e_82d8_164ae3a71623.slice/crio-2c96363dd88f6113a108e56efb09a3663af182ab9d5ec93b825f0b013a653b28 WatchSource:0}: Error finding container 2c96363dd88f6113a108e56efb09a3663af182ab9d5ec93b825f0b013a653b28: Status 404 returned error can't find the container with id 2c96363dd88f6113a108e56efb09a3663af182ab9d5ec93b825f0b013a653b28 Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.423191 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.465303 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.504456 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:10:33 crc kubenswrapper[4904]: E0214 11:10:33.504661 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:10:37.504627824 +0000 UTC m=+28.317392485 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.509212 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.546303 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.605447 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.605496 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.605549 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.605572 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:33 crc kubenswrapper[4904]: E0214 11:10:33.605614 4904 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 14 11:10:33 crc kubenswrapper[4904]: E0214 11:10:33.605648 4904 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 14 11:10:33 crc kubenswrapper[4904]: E0214 11:10:33.605668 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 14 11:10:33 crc kubenswrapper[4904]: E0214 11:10:33.605692 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 14 11:10:33 crc kubenswrapper[4904]: E0214 11:10:33.605702 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:37.60568273 +0000 UTC m=+28.418447461 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 14 11:10:33 crc kubenswrapper[4904]: E0214 11:10:33.605705 4904 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:33 crc kubenswrapper[4904]: E0214 11:10:33.605726 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:37.60571523 +0000 UTC m=+28.418479981 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 14 11:10:33 crc kubenswrapper[4904]: E0214 11:10:33.605733 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 14 11:10:33 crc kubenswrapper[4904]: E0214 11:10:33.605744 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 14 11:10:33 crc kubenswrapper[4904]: E0214 11:10:33.605745 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:37.605734981 +0000 UTC m=+28.418499642 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:33 crc kubenswrapper[4904]: E0214 11:10:33.605754 4904 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:33 crc kubenswrapper[4904]: E0214 11:10:33.605799 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:37.605777502 +0000 UTC m=+28.418542163 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.704222 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.708743 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.720538 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.721804 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.737996 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.754436 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.767148 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.784137 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.789193 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 16:45:46.17403165 +0000 UTC Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.806490 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.836000 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.836075 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:33 crc kubenswrapper[4904]: E0214 11:10:33.836146 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:10:33 crc kubenswrapper[4904]: E0214 11:10:33.836261 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.842782 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:33 crc kubenswrapper[4904]: E0214 11:10:33.843033 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.852887 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.889127 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.921308 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:33 crc kubenswrapper[4904]: I0214 11:10:33.963266 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:33Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.004723 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:34Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.041529 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:34Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.078980 4904 generic.go:334] "Generic (PLEG): container finished" podID="4edce510-9ef8-498a-8b5c-9b9eb41847eb" containerID="aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12" exitCode=0 Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.079084 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" event={"ID":"4edce510-9ef8-498a-8b5c-9b9eb41847eb","Type":"ContainerDied","Data":"aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12"} Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.080690 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zrwvd" event={"ID":"04ac70df-4f95-4a9e-82d8-164ae3a71623","Type":"ContainerStarted","Data":"c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e"} Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.080738 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zrwvd" event={"ID":"04ac70df-4f95-4a9e-82d8-164ae3a71623","Type":"ContainerStarted","Data":"2c96363dd88f6113a108e56efb09a3663af182ab9d5ec93b825f0b013a653b28"} Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.085244 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerStarted","Data":"2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27"} Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.085274 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerStarted","Data":"e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175"} Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.091514 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:34Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.122613 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:34Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.166822 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:34Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.207534 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:34Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.243200 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:34Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.282272 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:34Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.325012 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:34Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.369087 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:34Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.401978 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:34Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.452019 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:34Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.492918 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:34Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.530209 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:34Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.566309 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:34Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.606668 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:34Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.647746 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:34Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:34 crc kubenswrapper[4904]: I0214 11:10:34.790146 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 23:36:05.953980924 +0000 UTC Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.089276 4904 generic.go:334] "Generic (PLEG): container finished" podID="4edce510-9ef8-498a-8b5c-9b9eb41847eb" containerID="04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668" exitCode=0 Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.089317 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" event={"ID":"4edce510-9ef8-498a-8b5c-9b9eb41847eb","Type":"ContainerDied","Data":"04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668"} Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.108915 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:35Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.123040 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:35Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.153772 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:35Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.200277 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:35Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.236798 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:35Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.254558 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:35Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.263514 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:35Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.277862 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:35Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.287217 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:35Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.300310 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:35Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.310758 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:35Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.325867 4904 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.325987 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:35Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.327514 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.327546 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.327555 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.327628 4904 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.336781 4904 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.337115 4904 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.338223 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.338262 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.338271 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.338286 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.338294 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:35Z","lastTransitionTime":"2026-02-14T11:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.341120 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:35Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:35 crc kubenswrapper[4904]: E0214 11:10:35.361000 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:35Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.361717 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:35Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.364484 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.364504 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.364512 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.364524 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.364533 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:35Z","lastTransitionTime":"2026-02-14T11:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:35 crc kubenswrapper[4904]: E0214 11:10:35.374931 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:35Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.377764 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.377798 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.377812 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.377853 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.377867 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:35Z","lastTransitionTime":"2026-02-14T11:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:35 crc kubenswrapper[4904]: E0214 11:10:35.387650 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:35Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.390267 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.390287 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.390295 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.390307 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.390316 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:35Z","lastTransitionTime":"2026-02-14T11:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:35 crc kubenswrapper[4904]: E0214 11:10:35.399515 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:35Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.402870 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.402896 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.402907 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.402919 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.402927 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:35Z","lastTransitionTime":"2026-02-14T11:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:35 crc kubenswrapper[4904]: E0214 11:10:35.414106 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:35Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:35 crc kubenswrapper[4904]: E0214 11:10:35.414207 4904 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.415361 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.415383 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.415390 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.415402 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.415411 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:35Z","lastTransitionTime":"2026-02-14T11:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.517716 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.517750 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.517760 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.517774 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.517784 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:35Z","lastTransitionTime":"2026-02-14T11:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.619670 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.619700 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.619707 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.619720 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.619729 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:35Z","lastTransitionTime":"2026-02-14T11:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.722245 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.722275 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.722284 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.722300 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.722309 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:35Z","lastTransitionTime":"2026-02-14T11:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.790292 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 23:26:41.213931602 +0000 UTC Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.824115 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.824135 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.824143 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.824156 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.824165 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:35Z","lastTransitionTime":"2026-02-14T11:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.836085 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:35 crc kubenswrapper[4904]: E0214 11:10:35.836189 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.836469 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:35 crc kubenswrapper[4904]: E0214 11:10:35.839861 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.840014 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:35 crc kubenswrapper[4904]: E0214 11:10:35.840097 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.926234 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.926272 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.926284 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.926298 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:35 crc kubenswrapper[4904]: I0214 11:10:35.926306 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:35Z","lastTransitionTime":"2026-02-14T11:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.028627 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.028681 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.028694 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.028711 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.028724 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:36Z","lastTransitionTime":"2026-02-14T11:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.094230 4904 generic.go:334] "Generic (PLEG): container finished" podID="4edce510-9ef8-498a-8b5c-9b9eb41847eb" containerID="1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc" exitCode=0 Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.094310 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" event={"ID":"4edce510-9ef8-498a-8b5c-9b9eb41847eb","Type":"ContainerDied","Data":"1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc"} Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.100606 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerStarted","Data":"0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db"} Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.121672 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:36Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.132046 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.132087 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.132100 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.132120 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.132132 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:36Z","lastTransitionTime":"2026-02-14T11:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.138960 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:36Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.151612 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:36Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.167716 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:36Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.182581 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:36Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.201182 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:36Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.224343 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:36Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.235364 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.235401 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.235415 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.235448 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.235464 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:36Z","lastTransitionTime":"2026-02-14T11:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.235635 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:36Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.254808 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:36Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.269275 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:36Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.284050 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:36Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.299121 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:36Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.314385 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:36Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.330588 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:36Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.337741 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.337769 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.337779 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.337794 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.337807 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:36Z","lastTransitionTime":"2026-02-14T11:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.439690 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.439718 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.439726 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.439739 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.439750 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:36Z","lastTransitionTime":"2026-02-14T11:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.541590 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.541614 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.541622 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.541635 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.541655 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:36Z","lastTransitionTime":"2026-02-14T11:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.640789 4904 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.643499 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.643530 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.643541 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.643555 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.643594 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:36Z","lastTransitionTime":"2026-02-14T11:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.745951 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.746012 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.746034 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.746063 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.746084 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:36Z","lastTransitionTime":"2026-02-14T11:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.791110 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 02:43:19.051403641 +0000 UTC Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.848437 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.848470 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.848481 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.848497 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.848509 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:36Z","lastTransitionTime":"2026-02-14T11:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.952222 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.952264 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.952281 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.952305 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:36 crc kubenswrapper[4904]: I0214 11:10:36.952320 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:36Z","lastTransitionTime":"2026-02-14T11:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.055107 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.055136 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.055146 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.055161 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.055175 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:37Z","lastTransitionTime":"2026-02-14T11:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.113250 4904 generic.go:334] "Generic (PLEG): container finished" podID="4edce510-9ef8-498a-8b5c-9b9eb41847eb" containerID="406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a" exitCode=0 Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.113308 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" event={"ID":"4edce510-9ef8-498a-8b5c-9b9eb41847eb","Type":"ContainerDied","Data":"406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a"} Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.134013 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:37Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.151259 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:37Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.157861 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.157894 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.157906 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.157923 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.157936 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:37Z","lastTransitionTime":"2026-02-14T11:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.168396 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:37Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.187040 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:37Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.204977 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:37Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.225137 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:37Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.246633 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:37Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.257783 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:37Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.260591 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.260619 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.260630 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.260644 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.260654 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:37Z","lastTransitionTime":"2026-02-14T11:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.278466 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:37Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.292938 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:37Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.304213 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:37Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.322712 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:37Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.336391 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:37Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.348523 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:37Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.363016 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.363052 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.363062 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.363075 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.363085 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:37Z","lastTransitionTime":"2026-02-14T11:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.465676 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.465723 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.465735 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.465751 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.466123 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:37Z","lastTransitionTime":"2026-02-14T11:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.541265 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:10:37 crc kubenswrapper[4904]: E0214 11:10:37.542030 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:10:45.542005906 +0000 UTC m=+36.354770597 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.569735 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.569782 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.569799 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.569820 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.569867 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:37Z","lastTransitionTime":"2026-02-14T11:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.642986 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.643042 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.643074 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.643096 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:37 crc kubenswrapper[4904]: E0214 11:10:37.643341 4904 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 14 11:10:37 crc kubenswrapper[4904]: E0214 11:10:37.643411 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:45.64339604 +0000 UTC m=+36.456160701 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 14 11:10:37 crc kubenswrapper[4904]: E0214 11:10:37.643600 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 14 11:10:37 crc kubenswrapper[4904]: E0214 11:10:37.643682 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 14 11:10:37 crc kubenswrapper[4904]: E0214 11:10:37.643704 4904 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:37 crc kubenswrapper[4904]: E0214 11:10:37.643778 4904 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 14 11:10:37 crc kubenswrapper[4904]: E0214 11:10:37.643800 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:45.643767059 +0000 UTC m=+36.456531760 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:37 crc kubenswrapper[4904]: E0214 11:10:37.643841 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 14 11:10:37 crc kubenswrapper[4904]: E0214 11:10:37.643865 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 14 11:10:37 crc kubenswrapper[4904]: E0214 11:10:37.643877 4904 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:37 crc kubenswrapper[4904]: E0214 11:10:37.643886 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:45.64382318 +0000 UTC m=+36.456587961 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 14 11:10:37 crc kubenswrapper[4904]: E0214 11:10:37.643921 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-14 11:10:45.643910092 +0000 UTC m=+36.456674843 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.671757 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.671823 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.671897 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.671935 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.671957 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:37Z","lastTransitionTime":"2026-02-14T11:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.773803 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.773848 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.773858 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.773871 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.773879 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:37Z","lastTransitionTime":"2026-02-14T11:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.791344 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 23:24:49.060480192 +0000 UTC Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.836001 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:37 crc kubenswrapper[4904]: E0214 11:10:37.836205 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.836337 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:37 crc kubenswrapper[4904]: E0214 11:10:37.836625 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.837563 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:37 crc kubenswrapper[4904]: E0214 11:10:37.837673 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.875816 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.875863 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.875872 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.875886 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.875897 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:37Z","lastTransitionTime":"2026-02-14T11:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.978287 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.978335 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.978349 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.978367 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:37 crc kubenswrapper[4904]: I0214 11:10:37.978383 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:37Z","lastTransitionTime":"2026-02-14T11:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.080461 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.080500 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.080513 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.080531 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.080542 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:38Z","lastTransitionTime":"2026-02-14T11:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.118177 4904 generic.go:334] "Generic (PLEG): container finished" podID="4edce510-9ef8-498a-8b5c-9b9eb41847eb" containerID="e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86" exitCode=0 Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.118225 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" event={"ID":"4edce510-9ef8-498a-8b5c-9b9eb41847eb","Type":"ContainerDied","Data":"e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86"} Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.122895 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerStarted","Data":"443063beaf0f8c3a9321fe7252a89c7039fb0ca529508104aa2d279fcc699471"} Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.123242 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.123277 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.123286 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.172298 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.182982 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.183011 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.183021 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.183038 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.183049 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:38Z","lastTransitionTime":"2026-02-14T11:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.198965 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.213799 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.221166 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.223661 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.235261 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.247571 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.261714 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.271763 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.282284 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.287511 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.287553 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.287609 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.287632 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.287649 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:38Z","lastTransitionTime":"2026-02-14T11:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.311648 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.324224 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.337462 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.351428 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.364143 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.379821 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.389359 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.389928 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.389971 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.389980 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.389993 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.390001 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:38Z","lastTransitionTime":"2026-02-14T11:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.412876 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://443063beaf0f8c3a9321fe7252a89c7039fb0ca529508104aa2d279fcc699471\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.426529 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.441683 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.454584 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.467222 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.483182 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.492340 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.492380 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.492395 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.492416 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.492431 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:38Z","lastTransitionTime":"2026-02-14T11:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.497257 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.508737 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.509754 4904 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.519668 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.533208 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.544649 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.554605 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.567353 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:38Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.595344 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.595400 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.595419 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.595446 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.595463 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:38Z","lastTransitionTime":"2026-02-14T11:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.699140 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.699175 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.699183 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.699209 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.699219 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:38Z","lastTransitionTime":"2026-02-14T11:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.792388 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 00:08:20.450647615 +0000 UTC Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.801884 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.801938 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.801953 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.801973 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.801987 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:38Z","lastTransitionTime":"2026-02-14T11:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.904659 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.904689 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.904698 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.904712 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:38 crc kubenswrapper[4904]: I0214 11:10:38.904720 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:38Z","lastTransitionTime":"2026-02-14T11:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.007613 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.007683 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.007705 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.007770 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.007794 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:39Z","lastTransitionTime":"2026-02-14T11:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.110857 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.110899 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.110910 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.110926 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.110938 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:39Z","lastTransitionTime":"2026-02-14T11:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.129890 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" event={"ID":"4edce510-9ef8-498a-8b5c-9b9eb41847eb","Type":"ContainerStarted","Data":"709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593"} Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.144823 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.159581 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.176689 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.197419 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.211156 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.214367 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.214393 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.214402 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.214416 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.214425 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:39Z","lastTransitionTime":"2026-02-14T11:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.226521 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.236133 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.244619 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.261035 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://443063beaf0f8c3a9321fe7252a89c7039fb0ca529508104aa2d279fcc699471\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.273854 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.286543 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.297464 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.308328 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.316740 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.316766 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.316774 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.316787 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.316796 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:39Z","lastTransitionTime":"2026-02-14T11:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.322807 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.419091 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.419314 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.419408 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.419489 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.419560 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:39Z","lastTransitionTime":"2026-02-14T11:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.521878 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.522135 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.522286 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.522486 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.522657 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:39Z","lastTransitionTime":"2026-02-14T11:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.625164 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.625553 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.625705 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.625868 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.626008 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:39Z","lastTransitionTime":"2026-02-14T11:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.729053 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.729354 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.729616 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.729938 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.730194 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:39Z","lastTransitionTime":"2026-02-14T11:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.792973 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 13:04:10.223356269 +0000 UTC Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.832433 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.832633 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.832739 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.832823 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.832948 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:39Z","lastTransitionTime":"2026-02-14T11:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.835880 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:39 crc kubenswrapper[4904]: E0214 11:10:39.836409 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.836585 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:39 crc kubenswrapper[4904]: E0214 11:10:39.837079 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.836264 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:39 crc kubenswrapper[4904]: E0214 11:10:39.837363 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.860750 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.877306 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.888869 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.904755 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.927608 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.942353 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.942387 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.942397 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.942413 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.942424 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:39Z","lastTransitionTime":"2026-02-14T11:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.949241 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.975133 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:39 crc kubenswrapper[4904]: I0214 11:10:39.997033 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://443063beaf0f8c3a9321fe7252a89c7039fb0ca529508104aa2d279fcc699471\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:39Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.018280 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:40Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.030487 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:40Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.040950 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:40Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.044086 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.044120 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.044129 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.044142 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.044151 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:40Z","lastTransitionTime":"2026-02-14T11:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.052462 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:40Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.063297 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:40Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.074819 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:40Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.146229 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.146267 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.146278 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.146294 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.146306 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:40Z","lastTransitionTime":"2026-02-14T11:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.248261 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.248299 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.248309 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.248324 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.248336 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:40Z","lastTransitionTime":"2026-02-14T11:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.351250 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.351290 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.351298 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.351313 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.351331 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:40Z","lastTransitionTime":"2026-02-14T11:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.453346 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.453382 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.453390 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.453432 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.453442 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:40Z","lastTransitionTime":"2026-02-14T11:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.555165 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.555197 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.555205 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.555218 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.555228 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:40Z","lastTransitionTime":"2026-02-14T11:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.657150 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.657195 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.657204 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.657217 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.657226 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:40Z","lastTransitionTime":"2026-02-14T11:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.759185 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.759224 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.759234 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.759250 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.759262 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:40Z","lastTransitionTime":"2026-02-14T11:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.793987 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 20:43:28.05703832 +0000 UTC Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.861095 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.861131 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.861143 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.861158 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.861169 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:40Z","lastTransitionTime":"2026-02-14T11:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.963433 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.963467 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.963475 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.963488 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:40 crc kubenswrapper[4904]: I0214 11:10:40.963497 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:40Z","lastTransitionTime":"2026-02-14T11:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.065813 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.065884 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.065897 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.065917 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.065936 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:41Z","lastTransitionTime":"2026-02-14T11:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.137467 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovnkube-controller/0.log" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.141070 4904 generic.go:334] "Generic (PLEG): container finished" podID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerID="443063beaf0f8c3a9321fe7252a89c7039fb0ca529508104aa2d279fcc699471" exitCode=1 Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.141110 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerDied","Data":"443063beaf0f8c3a9321fe7252a89c7039fb0ca529508104aa2d279fcc699471"} Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.141676 4904 scope.go:117] "RemoveContainer" containerID="443063beaf0f8c3a9321fe7252a89c7039fb0ca529508104aa2d279fcc699471" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.155088 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:41Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.168054 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.168080 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.168089 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.168101 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.168112 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:41Z","lastTransitionTime":"2026-02-14T11:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.168598 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:41Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.180478 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:41Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.182147 4904 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.190333 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:41Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.204391 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:41Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.215356 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:41Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.227712 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:41Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.239974 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:41Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.250472 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:41Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.260380 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:41Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.270529 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.270578 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.270596 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.270618 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.270633 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:41Z","lastTransitionTime":"2026-02-14T11:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.275511 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:41Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.288667 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:41Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.299084 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:41Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.319724 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://443063beaf0f8c3a9321fe7252a89c7039fb0ca529508104aa2d279fcc699471\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://443063beaf0f8c3a9321fe7252a89c7039fb0ca529508104aa2d279fcc699471\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:40Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0214 11:10:40.559014 6127 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0214 11:10:40.561094 6127 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0214 11:10:40.561119 6127 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0214 11:10:40.561136 6127 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0214 11:10:40.561140 6127 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0214 11:10:40.561161 6127 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0214 11:10:40.561170 6127 handler.go:208] Removed *v1.Node event handler 7\\\\nI0214 11:10:40.561186 6127 handler.go:208] Removed *v1.Node event handler 2\\\\nI0214 11:10:40.561186 6127 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0214 11:10:40.561193 6127 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0214 11:10:40.561250 6127 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0214 11:10:40.561305 6127 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0214 11:10:40.561388 6127 factory.go:656] Stopping watch factory\\\\nI0214 11:10:40.561425 6127 ovnkube.go:599] Stopped ovnkube\\\\nI0214 11:10:40.561512 6127 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0214 11:10:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:41Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.373256 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.373282 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.373292 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.373304 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.373313 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:41Z","lastTransitionTime":"2026-02-14T11:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.477438 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.477486 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.477496 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.477510 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.477520 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:41Z","lastTransitionTime":"2026-02-14T11:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.579346 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.579379 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.579387 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.579400 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.579410 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:41Z","lastTransitionTime":"2026-02-14T11:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.681211 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.681510 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.681520 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.681533 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.681542 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:41Z","lastTransitionTime":"2026-02-14T11:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.783406 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.783443 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.783451 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.783469 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.783480 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:41Z","lastTransitionTime":"2026-02-14T11:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.795040 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 17:18:55.954430329 +0000 UTC Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.835873 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.835909 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.835982 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:41 crc kubenswrapper[4904]: E0214 11:10:41.836044 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:10:41 crc kubenswrapper[4904]: E0214 11:10:41.836114 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:10:41 crc kubenswrapper[4904]: E0214 11:10:41.836172 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.885760 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.885786 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.885794 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.885806 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.885815 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:41Z","lastTransitionTime":"2026-02-14T11:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.987484 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.987526 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.987538 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.987560 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:41 crc kubenswrapper[4904]: I0214 11:10:41.987577 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:41Z","lastTransitionTime":"2026-02-14T11:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.089506 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.089533 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.089541 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.089553 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.089561 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:42Z","lastTransitionTime":"2026-02-14T11:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.148675 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovnkube-controller/1.log" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.149738 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovnkube-controller/0.log" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.152796 4904 generic.go:334] "Generic (PLEG): container finished" podID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerID="3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4" exitCode=1 Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.152851 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerDied","Data":"3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4"} Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.152926 4904 scope.go:117] "RemoveContainer" containerID="443063beaf0f8c3a9321fe7252a89c7039fb0ca529508104aa2d279fcc699471" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.154067 4904 scope.go:117] "RemoveContainer" containerID="3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4" Feb 14 11:10:42 crc kubenswrapper[4904]: E0214 11:10:42.154485 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2gwlw_openshift-ovn-kubernetes(1ca5acc0-8b15-4052-883f-1987b7ca0067)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.167270 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:42Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.186674 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:42Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.191565 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.191584 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.191593 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.191605 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.191614 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:42Z","lastTransitionTime":"2026-02-14T11:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.200757 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:42Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.212811 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:42Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.223498 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:42Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.234977 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:42Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.247270 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:42Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.258759 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:42Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.270457 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:42Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.283404 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:42Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.293674 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.293718 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.293733 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.293751 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.293764 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:42Z","lastTransitionTime":"2026-02-14T11:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.293700 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:42Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.308212 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:42Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.317252 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:42Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.338257 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://443063beaf0f8c3a9321fe7252a89c7039fb0ca529508104aa2d279fcc699471\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:40Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0214 11:10:40.559014 6127 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0214 11:10:40.561094 6127 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0214 11:10:40.561119 6127 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0214 11:10:40.561136 6127 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0214 11:10:40.561140 6127 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0214 11:10:40.561161 6127 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0214 11:10:40.561170 6127 handler.go:208] Removed *v1.Node event handler 7\\\\nI0214 11:10:40.561186 6127 handler.go:208] Removed *v1.Node event handler 2\\\\nI0214 11:10:40.561186 6127 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0214 11:10:40.561193 6127 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0214 11:10:40.561250 6127 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0214 11:10:40.561305 6127 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0214 11:10:40.561388 6127 factory.go:656] Stopping watch factory\\\\nI0214 11:10:40.561425 6127 ovnkube.go:599] Stopped ovnkube\\\\nI0214 11:10:40.561512 6127 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0214 11:10:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:41Z\\\",\\\"message\\\":\\\"openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872464 6254 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872433 6254 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-jz5f5 in node crc\\\\nI0214 11:10:41.872524 6254 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-jz5f5 after 0 failed attempt(s)\\\\nI0214 11:10:41.872575 6254 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-jz5f5\\\\nF0214 11:10:41.872341 6254 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:42Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.395260 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.395292 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.395301 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.395314 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.395323 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:42Z","lastTransitionTime":"2026-02-14T11:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.497750 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.497791 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.497802 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.497819 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.497851 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:42Z","lastTransitionTime":"2026-02-14T11:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.599760 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.600100 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.600243 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.600373 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.600498 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:42Z","lastTransitionTime":"2026-02-14T11:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.703374 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.703414 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.703425 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.703441 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.703451 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:42Z","lastTransitionTime":"2026-02-14T11:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.795588 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 00:10:30.297416266 +0000 UTC Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.805352 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.805395 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.805405 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.805419 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.805428 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:42Z","lastTransitionTime":"2026-02-14T11:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.908567 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.908613 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.908625 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.908644 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:42 crc kubenswrapper[4904]: I0214 11:10:42.908659 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:42Z","lastTransitionTime":"2026-02-14T11:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.011201 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.011248 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.011261 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.011278 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.011290 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:43Z","lastTransitionTime":"2026-02-14T11:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.113240 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.113281 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.113292 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.113307 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.113319 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:43Z","lastTransitionTime":"2026-02-14T11:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.156858 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovnkube-controller/1.log" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.214982 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.215013 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.215023 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.215035 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.215043 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:43Z","lastTransitionTime":"2026-02-14T11:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.317129 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.317155 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.317165 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.317179 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.317189 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:43Z","lastTransitionTime":"2026-02-14T11:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.419590 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.419624 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.419635 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.419651 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.419662 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:43Z","lastTransitionTime":"2026-02-14T11:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.521480 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.521512 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.521521 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.521535 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.521544 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:43Z","lastTransitionTime":"2026-02-14T11:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.624609 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.624644 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.624652 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.624665 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.624673 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:43Z","lastTransitionTime":"2026-02-14T11:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.721902 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm"] Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.722484 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.724684 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.724751 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.727256 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.727288 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.727299 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.727313 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.727323 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:43Z","lastTransitionTime":"2026-02-14T11:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.737466 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:43Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.754315 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:43Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.770143 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:43Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.816374 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 18:46:55.402730521 +0000 UTC Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.819927 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f26819f5-4042-4943-b973-dba56706542e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-f5pqm\" (UID: \"f26819f5-4042-4943-b973-dba56706542e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.819966 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f26819f5-4042-4943-b973-dba56706542e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-f5pqm\" (UID: \"f26819f5-4042-4943-b973-dba56706542e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.820002 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt5kv\" (UniqueName: \"kubernetes.io/projected/f26819f5-4042-4943-b973-dba56706542e-kube-api-access-vt5kv\") pod \"ovnkube-control-plane-749d76644c-f5pqm\" (UID: \"f26819f5-4042-4943-b973-dba56706542e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.820027 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f26819f5-4042-4943-b973-dba56706542e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-f5pqm\" (UID: \"f26819f5-4042-4943-b973-dba56706542e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.822051 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:43Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.829517 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.829748 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.829883 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.830128 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.830354 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:43Z","lastTransitionTime":"2026-02-14T11:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.835760 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.835760 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:43 crc kubenswrapper[4904]: E0214 11:10:43.835874 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:10:43 crc kubenswrapper[4904]: E0214 11:10:43.835939 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.835773 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:43 crc kubenswrapper[4904]: E0214 11:10:43.836005 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.837257 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:43Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.854514 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:43Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.866587 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f26819f5-4042-4943-b973-dba56706542e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f5pqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:43Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.877129 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:43Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.888722 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:43Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.900756 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:43Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.910090 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:43Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.920469 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt5kv\" (UniqueName: \"kubernetes.io/projected/f26819f5-4042-4943-b973-dba56706542e-kube-api-access-vt5kv\") pod \"ovnkube-control-plane-749d76644c-f5pqm\" (UID: \"f26819f5-4042-4943-b973-dba56706542e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.920706 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f26819f5-4042-4943-b973-dba56706542e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-f5pqm\" (UID: \"f26819f5-4042-4943-b973-dba56706542e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.921172 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f26819f5-4042-4943-b973-dba56706542e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-f5pqm\" (UID: \"f26819f5-4042-4943-b973-dba56706542e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.921277 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f26819f5-4042-4943-b973-dba56706542e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-f5pqm\" (UID: \"f26819f5-4042-4943-b973-dba56706542e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.921298 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f26819f5-4042-4943-b973-dba56706542e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-f5pqm\" (UID: \"f26819f5-4042-4943-b973-dba56706542e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.921711 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f26819f5-4042-4943-b973-dba56706542e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-f5pqm\" (UID: \"f26819f5-4042-4943-b973-dba56706542e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.932129 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:43Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.933135 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.933151 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.933160 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.933172 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.933180 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:43Z","lastTransitionTime":"2026-02-14T11:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.933414 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f26819f5-4042-4943-b973-dba56706542e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-f5pqm\" (UID: \"f26819f5-4042-4943-b973-dba56706542e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.937694 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt5kv\" (UniqueName: \"kubernetes.io/projected/f26819f5-4042-4943-b973-dba56706542e-kube-api-access-vt5kv\") pod \"ovnkube-control-plane-749d76644c-f5pqm\" (UID: \"f26819f5-4042-4943-b973-dba56706542e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.942608 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:43Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.952779 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:43Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:43 crc kubenswrapper[4904]: I0214 11:10:43.969170 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://443063beaf0f8c3a9321fe7252a89c7039fb0ca529508104aa2d279fcc699471\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:40Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0214 11:10:40.559014 6127 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0214 11:10:40.561094 6127 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0214 11:10:40.561119 6127 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0214 11:10:40.561136 6127 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0214 11:10:40.561140 6127 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0214 11:10:40.561161 6127 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0214 11:10:40.561170 6127 handler.go:208] Removed *v1.Node event handler 7\\\\nI0214 11:10:40.561186 6127 handler.go:208] Removed *v1.Node event handler 2\\\\nI0214 11:10:40.561186 6127 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0214 11:10:40.561193 6127 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0214 11:10:40.561250 6127 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0214 11:10:40.561305 6127 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0214 11:10:40.561388 6127 factory.go:656] Stopping watch factory\\\\nI0214 11:10:40.561425 6127 ovnkube.go:599] Stopped ovnkube\\\\nI0214 11:10:40.561512 6127 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0214 11:10:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:41Z\\\",\\\"message\\\":\\\"openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872464 6254 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872433 6254 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-jz5f5 in node crc\\\\nI0214 11:10:41.872524 6254 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-jz5f5 after 0 failed attempt(s)\\\\nI0214 11:10:41.872575 6254 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-jz5f5\\\\nF0214 11:10:41.872341 6254 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:43Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.034467 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.034499 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.034509 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.034523 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.034533 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:44Z","lastTransitionTime":"2026-02-14T11:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.041573 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" Feb 14 11:10:44 crc kubenswrapper[4904]: W0214 11:10:44.051682 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf26819f5_4042_4943_b973_dba56706542e.slice/crio-181d45c5992203f7709d7d7f0e5e4080f1958db7e9447adc34d59a568d7999a4 WatchSource:0}: Error finding container 181d45c5992203f7709d7d7f0e5e4080f1958db7e9447adc34d59a568d7999a4: Status 404 returned error can't find the container with id 181d45c5992203f7709d7d7f0e5e4080f1958db7e9447adc34d59a568d7999a4 Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.137195 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.137231 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.137243 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.137261 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.137272 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:44Z","lastTransitionTime":"2026-02-14T11:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.163052 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" event={"ID":"f26819f5-4042-4943-b973-dba56706542e","Type":"ContainerStarted","Data":"181d45c5992203f7709d7d7f0e5e4080f1958db7e9447adc34d59a568d7999a4"} Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.239230 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.239264 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.239274 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.239290 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.239299 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:44Z","lastTransitionTime":"2026-02-14T11:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.341656 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.341709 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.341723 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.341743 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.341756 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:44Z","lastTransitionTime":"2026-02-14T11:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.444720 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.444754 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.444763 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.444778 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.444788 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:44Z","lastTransitionTime":"2026-02-14T11:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.546803 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.546862 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.546874 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.546892 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.546904 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:44Z","lastTransitionTime":"2026-02-14T11:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.649128 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.649164 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.649176 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.649190 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.649200 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:44Z","lastTransitionTime":"2026-02-14T11:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.751684 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.752057 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.752148 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.752661 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.752781 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:44Z","lastTransitionTime":"2026-02-14T11:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.816036 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-fz5f2"] Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.816727 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:10:44 crc kubenswrapper[4904]: E0214 11:10:44.816973 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.817097 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 23:30:46.044105877 +0000 UTC Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.828381 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs\") pod \"network-metrics-daemon-fz5f2\" (UID: \"ac91dd23-e4e5-4073-af76-d760dfdd1adc\") " pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.828461 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkv4r\" (UniqueName: \"kubernetes.io/projected/ac91dd23-e4e5-4073-af76-d760dfdd1adc-kube-api-access-zkv4r\") pod \"network-metrics-daemon-fz5f2\" (UID: \"ac91dd23-e4e5-4073-af76-d760dfdd1adc\") " pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.830330 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:44Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.841441 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:44Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.850596 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:44Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.855455 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.855506 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.855518 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.855532 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.855543 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:44Z","lastTransitionTime":"2026-02-14T11:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.862040 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:44Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.873811 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:44Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.883967 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:44Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.894582 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:44Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.915016 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://443063beaf0f8c3a9321fe7252a89c7039fb0ca529508104aa2d279fcc699471\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:40Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0214 11:10:40.559014 6127 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0214 11:10:40.561094 6127 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0214 11:10:40.561119 6127 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0214 11:10:40.561136 6127 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0214 11:10:40.561140 6127 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0214 11:10:40.561161 6127 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0214 11:10:40.561170 6127 handler.go:208] Removed *v1.Node event handler 7\\\\nI0214 11:10:40.561186 6127 handler.go:208] Removed *v1.Node event handler 2\\\\nI0214 11:10:40.561186 6127 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0214 11:10:40.561193 6127 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0214 11:10:40.561250 6127 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0214 11:10:40.561305 6127 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0214 11:10:40.561388 6127 factory.go:656] Stopping watch factory\\\\nI0214 11:10:40.561425 6127 ovnkube.go:599] Stopped ovnkube\\\\nI0214 11:10:40.561512 6127 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0214 11:10:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:41Z\\\",\\\"message\\\":\\\"openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872464 6254 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872433 6254 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-jz5f5 in node crc\\\\nI0214 11:10:41.872524 6254 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-jz5f5 after 0 failed attempt(s)\\\\nI0214 11:10:41.872575 6254 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-jz5f5\\\\nF0214 11:10:41.872341 6254 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:44Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.927980 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac91dd23-e4e5-4073-af76-d760dfdd1adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fz5f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:44Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.929265 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkv4r\" (UniqueName: \"kubernetes.io/projected/ac91dd23-e4e5-4073-af76-d760dfdd1adc-kube-api-access-zkv4r\") pod \"network-metrics-daemon-fz5f2\" (UID: \"ac91dd23-e4e5-4073-af76-d760dfdd1adc\") " pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.929321 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs\") pod \"network-metrics-daemon-fz5f2\" (UID: \"ac91dd23-e4e5-4073-af76-d760dfdd1adc\") " pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:10:44 crc kubenswrapper[4904]: E0214 11:10:44.929431 4904 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 14 11:10:44 crc kubenswrapper[4904]: E0214 11:10:44.929478 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs podName:ac91dd23-e4e5-4073-af76-d760dfdd1adc nodeName:}" failed. No retries permitted until 2026-02-14 11:10:45.429465314 +0000 UTC m=+36.242229975 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs") pod "network-metrics-daemon-fz5f2" (UID: "ac91dd23-e4e5-4073-af76-d760dfdd1adc") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.947589 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkv4r\" (UniqueName: \"kubernetes.io/projected/ac91dd23-e4e5-4073-af76-d760dfdd1adc-kube-api-access-zkv4r\") pod \"network-metrics-daemon-fz5f2\" (UID: \"ac91dd23-e4e5-4073-af76-d760dfdd1adc\") " pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.948796 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:44Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.957061 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.957104 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.957116 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.957132 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.957142 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:44Z","lastTransitionTime":"2026-02-14T11:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.963183 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:44Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.975041 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:44Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.986369 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:44Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:44 crc kubenswrapper[4904]: I0214 11:10:44.998877 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:44Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.013863 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.024137 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f26819f5-4042-4943-b973-dba56706542e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f5pqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.058791 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.058851 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.058863 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.058879 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.058890 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:45Z","lastTransitionTime":"2026-02-14T11:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.160758 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.160911 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.160931 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.160956 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.160973 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:45Z","lastTransitionTime":"2026-02-14T11:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.170647 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" event={"ID":"f26819f5-4042-4943-b973-dba56706542e","Type":"ContainerStarted","Data":"8ae046a4c4d0f2475bf13ed37f1f9c2029c5b73d436b31705e18ed030f7fc4b0"} Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.170685 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" event={"ID":"f26819f5-4042-4943-b973-dba56706542e","Type":"ContainerStarted","Data":"d7f9f319d3a100ef0c94c354ea3722a74d3541a399fe8ccedc4185d34934cbfd"} Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.185273 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.207140 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://443063beaf0f8c3a9321fe7252a89c7039fb0ca529508104aa2d279fcc699471\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:40Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0214 11:10:40.559014 6127 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0214 11:10:40.561094 6127 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0214 11:10:40.561119 6127 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0214 11:10:40.561136 6127 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0214 11:10:40.561140 6127 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0214 11:10:40.561161 6127 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0214 11:10:40.561170 6127 handler.go:208] Removed *v1.Node event handler 7\\\\nI0214 11:10:40.561186 6127 handler.go:208] Removed *v1.Node event handler 2\\\\nI0214 11:10:40.561186 6127 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0214 11:10:40.561193 6127 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0214 11:10:40.561250 6127 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0214 11:10:40.561305 6127 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0214 11:10:40.561388 6127 factory.go:656] Stopping watch factory\\\\nI0214 11:10:40.561425 6127 ovnkube.go:599] Stopped ovnkube\\\\nI0214 11:10:40.561512 6127 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0214 11:10:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:41Z\\\",\\\"message\\\":\\\"openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872464 6254 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872433 6254 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-jz5f5 in node crc\\\\nI0214 11:10:41.872524 6254 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-jz5f5 after 0 failed attempt(s)\\\\nI0214 11:10:41.872575 6254 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-jz5f5\\\\nF0214 11:10:41.872341 6254 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.220949 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac91dd23-e4e5-4073-af76-d760dfdd1adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fz5f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.234045 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.249609 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.263944 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.264034 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.264090 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.264117 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.264137 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:45Z","lastTransitionTime":"2026-02-14T11:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.270640 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.293388 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.313886 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.329627 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.343860 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f26819f5-4042-4943-b973-dba56706542e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7f9f319d3a100ef0c94c354ea3722a74d3541a399fe8ccedc4185d34934cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae046a4c4d0f2475bf13ed37f1f9c2029c5b73d436b31705e18ed030f7fc4b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f5pqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.360573 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.366210 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.366249 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.366265 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.366287 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.366303 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:45Z","lastTransitionTime":"2026-02-14T11:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.374487 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.385930 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.396270 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.409608 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.419359 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.435018 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs\") pod \"network-metrics-daemon-fz5f2\" (UID: \"ac91dd23-e4e5-4073-af76-d760dfdd1adc\") " pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.435184 4904 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.435241 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs podName:ac91dd23-e4e5-4073-af76-d760dfdd1adc nodeName:}" failed. No retries permitted until 2026-02-14 11:10:46.435224635 +0000 UTC m=+37.247989306 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs") pod "network-metrics-daemon-fz5f2" (UID: "ac91dd23-e4e5-4073-af76-d760dfdd1adc") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.468800 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.468932 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.468993 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.469018 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.469034 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:45Z","lastTransitionTime":"2026-02-14T11:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.571945 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.572330 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.572486 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.572623 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.572772 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:45Z","lastTransitionTime":"2026-02-14T11:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.636984 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.637140 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:11:01.637120511 +0000 UTC m=+52.449885182 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.675434 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.675460 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.675468 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.675480 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.675489 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:45Z","lastTransitionTime":"2026-02-14T11:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.738151 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.738194 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.738210 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.738236 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.738305 4904 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.738346 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-14 11:11:01.738334451 +0000 UTC m=+52.551099112 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.738659 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.738678 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.738687 4904 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.738721 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-14 11:11:01.73871202 +0000 UTC m=+52.551476681 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.738788 4904 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.738921 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-14 11:11:01.738899744 +0000 UTC m=+52.551664475 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.738994 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.739077 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.739144 4904 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.739250 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-14 11:11:01.739235152 +0000 UTC m=+52.551999813 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.778261 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.778476 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.778556 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.778667 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.778749 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:45Z","lastTransitionTime":"2026-02-14T11:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.794540 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.794632 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.794681 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.794705 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.794721 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:45Z","lastTransitionTime":"2026-02-14T11:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.810141 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.814053 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.814213 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.814299 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.814365 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.814421 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:45Z","lastTransitionTime":"2026-02-14T11:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.817509 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 20:33:27.054779205 +0000 UTC Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.825231 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.829210 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.829260 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.829275 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.829296 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.829312 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:45Z","lastTransitionTime":"2026-02-14T11:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.835451 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.835556 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.835856 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.835952 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.836084 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.836135 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.843327 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.847088 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.847112 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.847122 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.847135 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.847161 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:45Z","lastTransitionTime":"2026-02-14T11:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.858293 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.861816 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.861868 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.861879 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.861890 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.861898 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:45Z","lastTransitionTime":"2026-02-14T11:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.874632 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:45Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:45 crc kubenswrapper[4904]: E0214 11:10:45.874759 4904 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.880920 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.880949 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.880958 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.880970 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.880979 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:45Z","lastTransitionTime":"2026-02-14T11:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.983131 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.983165 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.983175 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.983190 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:45 crc kubenswrapper[4904]: I0214 11:10:45.983200 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:45Z","lastTransitionTime":"2026-02-14T11:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.085223 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.085256 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.085266 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.085295 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.085315 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:46Z","lastTransitionTime":"2026-02-14T11:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.187337 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.187365 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.187376 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.187390 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.187400 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:46Z","lastTransitionTime":"2026-02-14T11:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.289334 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.289358 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.289367 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.289382 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.289391 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:46Z","lastTransitionTime":"2026-02-14T11:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.391607 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.391640 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.391649 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.391661 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.391670 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:46Z","lastTransitionTime":"2026-02-14T11:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.443147 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs\") pod \"network-metrics-daemon-fz5f2\" (UID: \"ac91dd23-e4e5-4073-af76-d760dfdd1adc\") " pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:10:46 crc kubenswrapper[4904]: E0214 11:10:46.443407 4904 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 14 11:10:46 crc kubenswrapper[4904]: E0214 11:10:46.443490 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs podName:ac91dd23-e4e5-4073-af76-d760dfdd1adc nodeName:}" failed. No retries permitted until 2026-02-14 11:10:48.443466739 +0000 UTC m=+39.256231440 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs") pod "network-metrics-daemon-fz5f2" (UID: "ac91dd23-e4e5-4073-af76-d760dfdd1adc") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.494491 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.494545 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.494556 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.494569 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.494579 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:46Z","lastTransitionTime":"2026-02-14T11:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.596959 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.597317 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.597449 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.597579 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.597801 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:46Z","lastTransitionTime":"2026-02-14T11:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.700503 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.700568 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.700594 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.700667 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.700692 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:46Z","lastTransitionTime":"2026-02-14T11:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.803481 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.803783 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.803936 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.804038 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.804119 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:46Z","lastTransitionTime":"2026-02-14T11:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.817924 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 08:56:08.001597837 +0000 UTC Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.835277 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:10:46 crc kubenswrapper[4904]: E0214 11:10:46.835390 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.907695 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.907779 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.907799 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.908407 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:46 crc kubenswrapper[4904]: I0214 11:10:46.908653 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:46Z","lastTransitionTime":"2026-02-14T11:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.012509 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.012827 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.013025 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.013223 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.013475 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:47Z","lastTransitionTime":"2026-02-14T11:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.116740 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.116781 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.116792 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.116810 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.116821 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:47Z","lastTransitionTime":"2026-02-14T11:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.219582 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.219615 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.219625 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.219642 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.219655 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:47Z","lastTransitionTime":"2026-02-14T11:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.322195 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.322228 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.322239 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.322255 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.322265 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:47Z","lastTransitionTime":"2026-02-14T11:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.424997 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.425037 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.425048 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.425085 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.425096 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:47Z","lastTransitionTime":"2026-02-14T11:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.527890 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.528103 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.528192 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.528314 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.528416 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:47Z","lastTransitionTime":"2026-02-14T11:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.630650 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.630718 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.630734 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.630799 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.630811 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:47Z","lastTransitionTime":"2026-02-14T11:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.733650 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.733685 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.733700 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.733718 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.733734 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:47Z","lastTransitionTime":"2026-02-14T11:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.818742 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 16:50:13.797361532 +0000 UTC Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.835520 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.835540 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:47 crc kubenswrapper[4904]: E0214 11:10:47.835663 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.835727 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:47 crc kubenswrapper[4904]: E0214 11:10:47.835807 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:10:47 crc kubenswrapper[4904]: E0214 11:10:47.835974 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.836110 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.836180 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.836199 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.836230 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.836248 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:47Z","lastTransitionTime":"2026-02-14T11:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.943566 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.943605 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.943617 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.943634 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:47 crc kubenswrapper[4904]: I0214 11:10:47.943646 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:47Z","lastTransitionTime":"2026-02-14T11:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.045984 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.046042 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.046062 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.046096 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.046113 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:48Z","lastTransitionTime":"2026-02-14T11:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.148194 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.148227 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.148262 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.148274 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.148285 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:48Z","lastTransitionTime":"2026-02-14T11:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.250480 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.250522 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.250536 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.250553 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.250565 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:48Z","lastTransitionTime":"2026-02-14T11:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.353182 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.353224 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.353236 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.353255 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.353266 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:48Z","lastTransitionTime":"2026-02-14T11:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.456688 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.456763 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.456783 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.456809 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.456867 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:48Z","lastTransitionTime":"2026-02-14T11:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.485247 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs\") pod \"network-metrics-daemon-fz5f2\" (UID: \"ac91dd23-e4e5-4073-af76-d760dfdd1adc\") " pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:10:48 crc kubenswrapper[4904]: E0214 11:10:48.485438 4904 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 14 11:10:48 crc kubenswrapper[4904]: E0214 11:10:48.485560 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs podName:ac91dd23-e4e5-4073-af76-d760dfdd1adc nodeName:}" failed. No retries permitted until 2026-02-14 11:10:52.485530681 +0000 UTC m=+43.298295382 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs") pod "network-metrics-daemon-fz5f2" (UID: "ac91dd23-e4e5-4073-af76-d760dfdd1adc") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.559744 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.559847 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.559860 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.559876 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.559890 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:48Z","lastTransitionTime":"2026-02-14T11:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.662218 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.662270 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.662287 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.662312 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.662330 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:48Z","lastTransitionTime":"2026-02-14T11:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.764981 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.765042 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.765103 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.765160 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.765184 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:48Z","lastTransitionTime":"2026-02-14T11:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.819253 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 21:43:17.175790824 +0000 UTC Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.819396 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.831962 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:48Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.835749 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:10:48 crc kubenswrapper[4904]: E0214 11:10:48.835897 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.853265 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://443063beaf0f8c3a9321fe7252a89c7039fb0ca529508104aa2d279fcc699471\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:40Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0214 11:10:40.559014 6127 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0214 11:10:40.561094 6127 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0214 11:10:40.561119 6127 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0214 11:10:40.561136 6127 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0214 11:10:40.561140 6127 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0214 11:10:40.561161 6127 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0214 11:10:40.561170 6127 handler.go:208] Removed *v1.Node event handler 7\\\\nI0214 11:10:40.561186 6127 handler.go:208] Removed *v1.Node event handler 2\\\\nI0214 11:10:40.561186 6127 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0214 11:10:40.561193 6127 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0214 11:10:40.561250 6127 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0214 11:10:40.561305 6127 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0214 11:10:40.561388 6127 factory.go:656] Stopping watch factory\\\\nI0214 11:10:40.561425 6127 ovnkube.go:599] Stopped ovnkube\\\\nI0214 11:10:40.561512 6127 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0214 11:10:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:41Z\\\",\\\"message\\\":\\\"openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872464 6254 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872433 6254 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-jz5f5 in node crc\\\\nI0214 11:10:41.872524 6254 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-jz5f5 after 0 failed attempt(s)\\\\nI0214 11:10:41.872575 6254 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-jz5f5\\\\nF0214 11:10:41.872341 6254 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:48Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.868241 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.868283 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.868297 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.868315 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.868329 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:48Z","lastTransitionTime":"2026-02-14T11:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.868244 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac91dd23-e4e5-4073-af76-d760dfdd1adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fz5f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:48Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.895250 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:48Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.938564 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:48Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.954344 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:48Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.965497 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:48Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.970915 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.970960 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.970969 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.970982 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.970991 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:48Z","lastTransitionTime":"2026-02-14T11:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.977316 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:48Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.987809 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:48Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:48 crc kubenswrapper[4904]: I0214 11:10:48.997060 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f26819f5-4042-4943-b973-dba56706542e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7f9f319d3a100ef0c94c354ea3722a74d3541a399fe8ccedc4185d34934cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae046a4c4d0f2475bf13ed37f1f9c2029c5b73d436b31705e18ed030f7fc4b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f5pqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:48Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.007884 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:49Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.018709 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:49Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.029146 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:49Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.039058 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:49Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.050629 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:49Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.059323 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:49Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.072969 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.073007 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.073016 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.073031 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.073039 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:49Z","lastTransitionTime":"2026-02-14T11:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.175257 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.175295 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.175304 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.175320 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.175330 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:49Z","lastTransitionTime":"2026-02-14T11:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.277394 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.277436 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.277447 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.277464 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.277476 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:49Z","lastTransitionTime":"2026-02-14T11:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.379887 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.379933 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.379946 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.379966 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.379977 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:49Z","lastTransitionTime":"2026-02-14T11:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.482278 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.482342 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.482360 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.482384 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.482405 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:49Z","lastTransitionTime":"2026-02-14T11:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.584994 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.585058 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.585075 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.585100 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.585118 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:49Z","lastTransitionTime":"2026-02-14T11:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.687702 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.688072 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.688272 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.688451 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.688580 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:49Z","lastTransitionTime":"2026-02-14T11:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.790429 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.790477 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.790488 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.790503 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.790514 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:49Z","lastTransitionTime":"2026-02-14T11:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.820061 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 08:26:23.956979568 +0000 UTC Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.836147 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.836227 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.836282 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:49 crc kubenswrapper[4904]: E0214 11:10:49.836406 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:10:49 crc kubenswrapper[4904]: E0214 11:10:49.836633 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:10:49 crc kubenswrapper[4904]: E0214 11:10:49.836732 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.852406 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:49Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.869936 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:49Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.885427 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:49Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.892905 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.892961 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.892973 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.892987 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.893025 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:49Z","lastTransitionTime":"2026-02-14T11:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.899735 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:49Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.916449 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:49Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.936801 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:49Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.949342 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:49Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.978022 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://443063beaf0f8c3a9321fe7252a89c7039fb0ca529508104aa2d279fcc699471\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:40Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0214 11:10:40.559014 6127 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0214 11:10:40.561094 6127 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0214 11:10:40.561119 6127 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0214 11:10:40.561136 6127 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0214 11:10:40.561140 6127 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0214 11:10:40.561161 6127 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0214 11:10:40.561170 6127 handler.go:208] Removed *v1.Node event handler 7\\\\nI0214 11:10:40.561186 6127 handler.go:208] Removed *v1.Node event handler 2\\\\nI0214 11:10:40.561186 6127 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0214 11:10:40.561193 6127 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0214 11:10:40.561250 6127 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0214 11:10:40.561305 6127 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0214 11:10:40.561388 6127 factory.go:656] Stopping watch factory\\\\nI0214 11:10:40.561425 6127 ovnkube.go:599] Stopped ovnkube\\\\nI0214 11:10:40.561512 6127 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0214 11:10:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:41Z\\\",\\\"message\\\":\\\"openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872464 6254 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872433 6254 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-jz5f5 in node crc\\\\nI0214 11:10:41.872524 6254 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-jz5f5 after 0 failed attempt(s)\\\\nI0214 11:10:41.872575 6254 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-jz5f5\\\\nF0214 11:10:41.872341 6254 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:49Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.991246 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac91dd23-e4e5-4073-af76-d760dfdd1adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fz5f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:49Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.995209 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.995386 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.995490 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.995580 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:49 crc kubenswrapper[4904]: I0214 11:10:49.995667 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:49Z","lastTransitionTime":"2026-02-14T11:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.008693 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:50Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.025410 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:50Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.043756 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:50Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.061373 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:50Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.077336 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:50Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.092220 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:50Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.097548 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.097605 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.097623 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.097648 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.097666 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:50Z","lastTransitionTime":"2026-02-14T11:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.107191 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f26819f5-4042-4943-b973-dba56706542e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7f9f319d3a100ef0c94c354ea3722a74d3541a399fe8ccedc4185d34934cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae046a4c4d0f2475bf13ed37f1f9c2029c5b73d436b31705e18ed030f7fc4b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f5pqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:50Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.199509 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.199551 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.199564 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.199582 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.199594 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:50Z","lastTransitionTime":"2026-02-14T11:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.302337 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.302442 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.302461 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.302520 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.302543 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:50Z","lastTransitionTime":"2026-02-14T11:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.405182 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.405237 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.405256 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.405280 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.405300 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:50Z","lastTransitionTime":"2026-02-14T11:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.508224 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.508282 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.508301 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.508326 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.508344 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:50Z","lastTransitionTime":"2026-02-14T11:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.611263 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.611331 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.611349 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.611374 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.611391 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:50Z","lastTransitionTime":"2026-02-14T11:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.716977 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.717006 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.717015 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.717035 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.717044 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:50Z","lastTransitionTime":"2026-02-14T11:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.819461 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.819512 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.819529 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.819553 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.819569 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:50Z","lastTransitionTime":"2026-02-14T11:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.820243 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 20:13:55.403276952 +0000 UTC Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.835683 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:10:50 crc kubenswrapper[4904]: E0214 11:10:50.835812 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.922956 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.923001 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.923014 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.923032 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:50 crc kubenswrapper[4904]: I0214 11:10:50.923050 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:50Z","lastTransitionTime":"2026-02-14T11:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.025812 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.025889 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.025903 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.025920 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.025932 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:51Z","lastTransitionTime":"2026-02-14T11:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.128109 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.128215 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.128233 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.128254 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.128278 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:51Z","lastTransitionTime":"2026-02-14T11:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.230777 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.230820 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.230849 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.230870 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.230882 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:51Z","lastTransitionTime":"2026-02-14T11:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.335155 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.335196 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.335208 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.335225 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.335237 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:51Z","lastTransitionTime":"2026-02-14T11:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.437522 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.437571 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.437581 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.437594 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.437604 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:51Z","lastTransitionTime":"2026-02-14T11:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.539812 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.540060 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.540124 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.540244 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.540325 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:51Z","lastTransitionTime":"2026-02-14T11:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.643290 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.643373 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.643386 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.643409 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.643424 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:51Z","lastTransitionTime":"2026-02-14T11:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.747304 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.747337 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.747351 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.747369 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.747380 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:51Z","lastTransitionTime":"2026-02-14T11:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.820706 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 15:30:57.71074687 +0000 UTC Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.835367 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:51 crc kubenswrapper[4904]: E0214 11:10:51.835536 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.836203 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.836219 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:51 crc kubenswrapper[4904]: E0214 11:10:51.836461 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:10:51 crc kubenswrapper[4904]: E0214 11:10:51.836588 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.850096 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.850137 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.850152 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.850172 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.850184 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:51Z","lastTransitionTime":"2026-02-14T11:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.953256 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.953310 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.953319 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.953335 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:51 crc kubenswrapper[4904]: I0214 11:10:51.953344 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:51Z","lastTransitionTime":"2026-02-14T11:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.056609 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.056638 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.056645 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.056660 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.056669 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:52Z","lastTransitionTime":"2026-02-14T11:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.159085 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.159114 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.159122 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.159133 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.159141 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:52Z","lastTransitionTime":"2026-02-14T11:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.261120 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.261161 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.261172 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.261197 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.261210 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:52Z","lastTransitionTime":"2026-02-14T11:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.364819 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.364908 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.364926 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.364952 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.364970 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:52Z","lastTransitionTime":"2026-02-14T11:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.467564 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.467622 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.467639 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.467660 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.467677 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:52Z","lastTransitionTime":"2026-02-14T11:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.535765 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs\") pod \"network-metrics-daemon-fz5f2\" (UID: \"ac91dd23-e4e5-4073-af76-d760dfdd1adc\") " pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:10:52 crc kubenswrapper[4904]: E0214 11:10:52.536046 4904 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 14 11:10:52 crc kubenswrapper[4904]: E0214 11:10:52.536171 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs podName:ac91dd23-e4e5-4073-af76-d760dfdd1adc nodeName:}" failed. No retries permitted until 2026-02-14 11:11:00.536129355 +0000 UTC m=+51.348894056 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs") pod "network-metrics-daemon-fz5f2" (UID: "ac91dd23-e4e5-4073-af76-d760dfdd1adc") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.571024 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.571179 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.571203 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.571237 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.571260 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:52Z","lastTransitionTime":"2026-02-14T11:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.674376 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.675045 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.675088 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.675119 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.675138 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:52Z","lastTransitionTime":"2026-02-14T11:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.778083 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.778138 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.778158 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.778187 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.778207 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:52Z","lastTransitionTime":"2026-02-14T11:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.821443 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 17:52:39.723941073 +0000 UTC Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.836033 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:10:52 crc kubenswrapper[4904]: E0214 11:10:52.836250 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.881757 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.881823 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.881879 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.881914 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.881934 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:52Z","lastTransitionTime":"2026-02-14T11:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.984887 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.984948 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.984965 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.985363 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:52 crc kubenswrapper[4904]: I0214 11:10:52.985381 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:52Z","lastTransitionTime":"2026-02-14T11:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.088421 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.088476 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.088489 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.088510 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.088520 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:53Z","lastTransitionTime":"2026-02-14T11:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.190794 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.190825 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.190846 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.190860 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.190869 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:53Z","lastTransitionTime":"2026-02-14T11:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.293625 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.293690 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.293706 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.293730 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.293754 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:53Z","lastTransitionTime":"2026-02-14T11:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.395781 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.395828 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.395870 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.395889 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.395905 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:53Z","lastTransitionTime":"2026-02-14T11:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.498211 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.498282 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.498305 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.498327 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.498344 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:53Z","lastTransitionTime":"2026-02-14T11:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.600409 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.600445 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.600456 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.600471 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.600482 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:53Z","lastTransitionTime":"2026-02-14T11:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.702774 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.702811 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.702822 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.702853 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.702863 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:53Z","lastTransitionTime":"2026-02-14T11:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.804857 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.804891 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.804902 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.804916 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.804928 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:53Z","lastTransitionTime":"2026-02-14T11:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.822194 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 01:54:25.520835479 +0000 UTC Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.835734 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.835733 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:53 crc kubenswrapper[4904]: E0214 11:10:53.835895 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:10:53 crc kubenswrapper[4904]: E0214 11:10:53.835910 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.836018 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:53 crc kubenswrapper[4904]: E0214 11:10:53.836187 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.907047 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.907086 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.907100 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.907116 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:53 crc kubenswrapper[4904]: I0214 11:10:53.907125 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:53Z","lastTransitionTime":"2026-02-14T11:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.414569 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.414620 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.414638 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.414661 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.414680 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:54Z","lastTransitionTime":"2026-02-14T11:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.517046 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.517086 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.517097 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.517115 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.517127 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:54Z","lastTransitionTime":"2026-02-14T11:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.619777 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.619881 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.619894 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.619910 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.619922 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:54Z","lastTransitionTime":"2026-02-14T11:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.721644 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.721681 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.721692 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.721708 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.721720 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:54Z","lastTransitionTime":"2026-02-14T11:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.822667 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 17:57:35.019489237 +0000 UTC Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.824185 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.824227 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.824236 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.824250 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.824259 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:54Z","lastTransitionTime":"2026-02-14T11:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.835606 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:10:54 crc kubenswrapper[4904]: E0214 11:10:54.835883 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.927399 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.927456 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.927473 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.927496 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:54 crc kubenswrapper[4904]: I0214 11:10:54.927515 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:54Z","lastTransitionTime":"2026-02-14T11:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.029473 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.029528 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.029546 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.029570 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.029588 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:55Z","lastTransitionTime":"2026-02-14T11:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.133443 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.133546 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.133565 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.133631 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.133662 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:55Z","lastTransitionTime":"2026-02-14T11:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.236717 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.236755 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.236766 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.236781 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.236793 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:55Z","lastTransitionTime":"2026-02-14T11:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.339864 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.339911 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.339929 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.339952 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.339972 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:55Z","lastTransitionTime":"2026-02-14T11:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.443297 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.443350 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.443362 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.443380 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.443392 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:55Z","lastTransitionTime":"2026-02-14T11:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.546620 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.546690 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.546725 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.546757 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.546778 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:55Z","lastTransitionTime":"2026-02-14T11:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.648692 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.648729 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.648741 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.648756 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.648767 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:55Z","lastTransitionTime":"2026-02-14T11:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.752908 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.752977 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.752999 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.753026 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.753047 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:55Z","lastTransitionTime":"2026-02-14T11:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.823775 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 17:17:02.89799408 +0000 UTC Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.836367 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.836426 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:55 crc kubenswrapper[4904]: E0214 11:10:55.836574 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.836726 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:55 crc kubenswrapper[4904]: E0214 11:10:55.836854 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:10:55 crc kubenswrapper[4904]: E0214 11:10:55.836923 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.855743 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.856037 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.856078 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.856114 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.856141 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:55Z","lastTransitionTime":"2026-02-14T11:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.958431 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.958476 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.958487 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.958505 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.958518 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:55Z","lastTransitionTime":"2026-02-14T11:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.992103 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.992413 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.992513 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.992601 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:55 crc kubenswrapper[4904]: I0214 11:10:55.992676 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:55Z","lastTransitionTime":"2026-02-14T11:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:56 crc kubenswrapper[4904]: E0214 11:10:56.010171 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:56Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.014598 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.014761 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.014906 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.015045 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.015125 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:56Z","lastTransitionTime":"2026-02-14T11:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:56 crc kubenswrapper[4904]: E0214 11:10:56.031949 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:56Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.036270 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.036320 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.036340 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.036395 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.036414 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:56Z","lastTransitionTime":"2026-02-14T11:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:56 crc kubenswrapper[4904]: E0214 11:10:56.053472 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:56Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.058523 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.058566 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.058580 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.058599 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.058612 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:56Z","lastTransitionTime":"2026-02-14T11:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:56 crc kubenswrapper[4904]: E0214 11:10:56.073024 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:56Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.077084 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.077189 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.077206 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.077224 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.077268 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:56Z","lastTransitionTime":"2026-02-14T11:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:56 crc kubenswrapper[4904]: E0214 11:10:56.096464 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:56Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:56 crc kubenswrapper[4904]: E0214 11:10:56.096655 4904 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.098529 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.098668 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.098782 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.098939 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.099045 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:56Z","lastTransitionTime":"2026-02-14T11:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.202229 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.202278 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.202291 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.202309 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.202321 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:56Z","lastTransitionTime":"2026-02-14T11:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.306292 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.306358 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.306401 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.306436 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.306462 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:56Z","lastTransitionTime":"2026-02-14T11:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.409274 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.409301 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.409310 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.409325 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.409333 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:56Z","lastTransitionTime":"2026-02-14T11:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.511627 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.511658 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.511666 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.511679 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.511687 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:56Z","lastTransitionTime":"2026-02-14T11:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.613631 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.613659 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.613669 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.613856 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.613868 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:56Z","lastTransitionTime":"2026-02-14T11:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.716444 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.716478 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.716488 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.716503 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.716513 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:56Z","lastTransitionTime":"2026-02-14T11:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.818630 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.818667 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.818675 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.818694 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.818705 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:56Z","lastTransitionTime":"2026-02-14T11:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.824090 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 22:57:48.754668303 +0000 UTC Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.836370 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.836969 4904 scope.go:117] "RemoveContainer" containerID="3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4" Feb 14 11:10:56 crc kubenswrapper[4904]: E0214 11:10:56.837307 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.856903 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:56Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.873136 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:56Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.885150 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:56Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.901253 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:56Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.914172 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:56Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.921977 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.922023 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.922039 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.922059 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.922073 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:56Z","lastTransitionTime":"2026-02-14T11:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.927636 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:56Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.939382 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:56Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.958848 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:41Z\\\",\\\"message\\\":\\\"openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872464 6254 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872433 6254 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-jz5f5 in node crc\\\\nI0214 11:10:41.872524 6254 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-jz5f5 after 0 failed attempt(s)\\\\nI0214 11:10:41.872575 6254 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-jz5f5\\\\nF0214 11:10:41.872341 6254 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2gwlw_openshift-ovn-kubernetes(1ca5acc0-8b15-4052-883f-1987b7ca0067)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:56Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.968961 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac91dd23-e4e5-4073-af76-d760dfdd1adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fz5f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:56Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:56 crc kubenswrapper[4904]: I0214 11:10:56.987023 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:56Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.002566 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:56Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.015974 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.024373 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.024399 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.024406 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.024419 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.024428 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:57Z","lastTransitionTime":"2026-02-14T11:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.028603 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.040369 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.049229 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f26819f5-4042-4943-b973-dba56706542e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7f9f319d3a100ef0c94c354ea3722a74d3541a399fe8ccedc4185d34934cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae046a4c4d0f2475bf13ed37f1f9c2029c5b73d436b31705e18ed030f7fc4b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f5pqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.058568 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.126405 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.126450 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.126465 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.126485 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.126503 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:57Z","lastTransitionTime":"2026-02-14T11:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.228437 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.228482 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.228494 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.228511 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.228522 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:57Z","lastTransitionTime":"2026-02-14T11:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.322719 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.330912 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.330936 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.330943 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.330954 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.330963 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:57Z","lastTransitionTime":"2026-02-14T11:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.333056 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.338347 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.347737 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.357623 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.366787 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.376826 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.389350 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.401133 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.420691 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:41Z\\\",\\\"message\\\":\\\"openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872464 6254 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872433 6254 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-jz5f5 in node crc\\\\nI0214 11:10:41.872524 6254 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-jz5f5 after 0 failed attempt(s)\\\\nI0214 11:10:41.872575 6254 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-jz5f5\\\\nF0214 11:10:41.872341 6254 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2gwlw_openshift-ovn-kubernetes(1ca5acc0-8b15-4052-883f-1987b7ca0067)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.431022 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac91dd23-e4e5-4073-af76-d760dfdd1adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fz5f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.432039 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovnkube-controller/1.log" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.432239 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.432260 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.432269 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.432281 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.432289 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:57Z","lastTransitionTime":"2026-02-14T11:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.435284 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerStarted","Data":"f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8"} Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.435809 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.443977 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.457929 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.472373 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.484716 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.494670 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.509470 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.519980 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f26819f5-4042-4943-b973-dba56706542e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7f9f319d3a100ef0c94c354ea3722a74d3541a399fe8ccedc4185d34934cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae046a4c4d0f2475bf13ed37f1f9c2029c5b73d436b31705e18ed030f7fc4b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f5pqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.529524 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.534392 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.534443 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.534459 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.534474 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.534483 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:57Z","lastTransitionTime":"2026-02-14T11:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.549410 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:41Z\\\",\\\"message\\\":\\\"openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872464 6254 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872433 6254 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-jz5f5 in node crc\\\\nI0214 11:10:41.872524 6254 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-jz5f5 after 0 failed attempt(s)\\\\nI0214 11:10:41.872575 6254 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-jz5f5\\\\nF0214 11:10:41.872341 6254 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.559343 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac91dd23-e4e5-4073-af76-d760dfdd1adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fz5f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.574609 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.587702 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.599005 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfcec17d-4d9a-4cdc-a86c-d05e21bedefb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee16aacae98d845ea68a860c26436a43d376f74ea7abd4396984fa63a836cb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc073ab93f3609414e270a023f8d2bad18f489eb6b37ac9a15cd6bf8b4dfbbf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2411332719f59b693409f1e2c9a10c280a102c11216e67d8eea8bfa8bdfe809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.608537 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.621315 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.637088 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.637131 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.637145 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.637167 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.637182 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:57Z","lastTransitionTime":"2026-02-14T11:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.637095 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.651535 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.664795 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f26819f5-4042-4943-b973-dba56706542e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7f9f319d3a100ef0c94c354ea3722a74d3541a399fe8ccedc4185d34934cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae046a4c4d0f2475bf13ed37f1f9c2029c5b73d436b31705e18ed030f7fc4b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f5pqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.678169 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.691968 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.704145 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.725091 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.739691 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.740470 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.740509 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.740526 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.740559 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.740569 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:57Z","lastTransitionTime":"2026-02-14T11:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.757239 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:57Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.824724 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 19:49:20.1396027 +0000 UTC Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.836292 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.836407 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.836354 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:57 crc kubenswrapper[4904]: E0214 11:10:57.836477 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:10:57 crc kubenswrapper[4904]: E0214 11:10:57.836660 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:10:57 crc kubenswrapper[4904]: E0214 11:10:57.836768 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.842965 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.843006 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.843016 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.843032 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.843042 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:57Z","lastTransitionTime":"2026-02-14T11:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.946273 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.946303 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.946312 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.946325 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:57 crc kubenswrapper[4904]: I0214 11:10:57.946333 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:57Z","lastTransitionTime":"2026-02-14T11:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.049459 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.049503 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.049514 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.049530 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.049540 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:58Z","lastTransitionTime":"2026-02-14T11:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.153620 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.153676 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.153696 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.153732 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.153745 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:58Z","lastTransitionTime":"2026-02-14T11:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.256271 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.256303 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.256317 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.256333 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.256345 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:58Z","lastTransitionTime":"2026-02-14T11:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.361208 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.361268 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.361282 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.361301 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.361314 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:58Z","lastTransitionTime":"2026-02-14T11:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.440685 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovnkube-controller/2.log" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.441919 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovnkube-controller/1.log" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.446538 4904 generic.go:334] "Generic (PLEG): container finished" podID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerID="f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8" exitCode=1 Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.446615 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerDied","Data":"f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8"} Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.446676 4904 scope.go:117] "RemoveContainer" containerID="3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.447406 4904 scope.go:117] "RemoveContainer" containerID="f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8" Feb 14 11:10:58 crc kubenswrapper[4904]: E0214 11:10:58.447607 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2gwlw_openshift-ovn-kubernetes(1ca5acc0-8b15-4052-883f-1987b7ca0067)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.464694 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:58Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.465450 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.465469 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.465478 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.465494 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.465504 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:58Z","lastTransitionTime":"2026-02-14T11:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.481358 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f26819f5-4042-4943-b973-dba56706542e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7f9f319d3a100ef0c94c354ea3722a74d3541a399fe8ccedc4185d34934cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae046a4c4d0f2475bf13ed37f1f9c2029c5b73d436b31705e18ed030f7fc4b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f5pqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:58Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.494272 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:58Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.509236 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:58Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.525065 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:58Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.544818 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:58Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.561177 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:58Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.568981 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.569041 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.569081 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.569107 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.569122 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:58Z","lastTransitionTime":"2026-02-14T11:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.575434 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:58Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.591945 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:58Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.613723 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a57b6c4d9234d245b096d9e51a90208f7c455a6acb4c052e511de43c9e114f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:41Z\\\",\\\"message\\\":\\\"openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872464 6254 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:10:41.872433 6254 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-jz5f5 in node crc\\\\nI0214 11:10:41.872524 6254 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-jz5f5 after 0 failed attempt(s)\\\\nI0214 11:10:41.872575 6254 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-jz5f5\\\\nF0214 11:10:41.872341 6254 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"message\\\":\\\"ift-kube-scheduler-crc openshift-multus/network-metrics-daemon-fz5f2 openshift-network-operator/iptables-alerter-4ln5h openshift-network-diagnostics/network-check-target-xd92c openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI0214 11:10:57.697651 6461 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI0214 11:10:57.697673 6461 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697684 6461 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697702 6461 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0214 11:10:57.697710 6461 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0214 11:10:57.697715 6461 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697733 6461 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0214 11:10:57.697811 6461 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:58Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.631123 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac91dd23-e4e5-4073-af76-d760dfdd1adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fz5f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:58Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.649722 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:58Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.667046 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:58Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.672546 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.672595 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.672615 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.672647 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.672671 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:58Z","lastTransitionTime":"2026-02-14T11:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.684363 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfcec17d-4d9a-4cdc-a86c-d05e21bedefb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee16aacae98d845ea68a860c26436a43d376f74ea7abd4396984fa63a836cb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc073ab93f3609414e270a023f8d2bad18f489eb6b37ac9a15cd6bf8b4dfbbf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2411332719f59b693409f1e2c9a10c280a102c11216e67d8eea8bfa8bdfe809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:58Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.701479 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:58Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.717260 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:58Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.742930 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:58Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.775394 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.775430 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.775441 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.775461 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.775474 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:58Z","lastTransitionTime":"2026-02-14T11:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.825810 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 12:07:36.159664872 +0000 UTC Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.836487 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:10:58 crc kubenswrapper[4904]: E0214 11:10:58.837091 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.879807 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.879870 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.879880 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.879896 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.879905 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:58Z","lastTransitionTime":"2026-02-14T11:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.983594 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.983637 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.983647 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.983664 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:58 crc kubenswrapper[4904]: I0214 11:10:58.983675 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:58Z","lastTransitionTime":"2026-02-14T11:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.087274 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.087328 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.087337 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.087356 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.087367 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:59Z","lastTransitionTime":"2026-02-14T11:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.189959 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.190041 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.190064 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.190090 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.190106 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:59Z","lastTransitionTime":"2026-02-14T11:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.294062 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.294124 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.294142 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.294167 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.294189 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:59Z","lastTransitionTime":"2026-02-14T11:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.396786 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.396917 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.396948 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.396988 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.397012 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:59Z","lastTransitionTime":"2026-02-14T11:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.454639 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovnkube-controller/2.log" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.459774 4904 scope.go:117] "RemoveContainer" containerID="f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8" Feb 14 11:10:59 crc kubenswrapper[4904]: E0214 11:10:59.460132 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2gwlw_openshift-ovn-kubernetes(1ca5acc0-8b15-4052-883f-1987b7ca0067)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.483054 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.500862 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.500982 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.501010 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.501021 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.501040 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.501052 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:59Z","lastTransitionTime":"2026-02-14T11:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.512115 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfcec17d-4d9a-4cdc-a86c-d05e21bedefb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee16aacae98d845ea68a860c26436a43d376f74ea7abd4396984fa63a836cb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc073ab93f3609414e270a023f8d2bad18f489eb6b37ac9a15cd6bf8b4dfbbf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2411332719f59b693409f1e2c9a10c280a102c11216e67d8eea8bfa8bdfe809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.529091 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.545589 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.574631 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.591363 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.603770 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.603891 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.603921 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.603955 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.603981 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:59Z","lastTransitionTime":"2026-02-14T11:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.608128 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f26819f5-4042-4943-b973-dba56706542e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7f9f319d3a100ef0c94c354ea3722a74d3541a399fe8ccedc4185d34934cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae046a4c4d0f2475bf13ed37f1f9c2029c5b73d436b31705e18ed030f7fc4b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f5pqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.626080 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.640606 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.657444 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.672525 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.695152 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.706495 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.706564 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.706576 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.706614 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.706627 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:59Z","lastTransitionTime":"2026-02-14T11:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.709983 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.724433 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.743676 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"message\\\":\\\"ift-kube-scheduler-crc openshift-multus/network-metrics-daemon-fz5f2 openshift-network-operator/iptables-alerter-4ln5h openshift-network-diagnostics/network-check-target-xd92c openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI0214 11:10:57.697651 6461 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI0214 11:10:57.697673 6461 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697684 6461 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697702 6461 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0214 11:10:57.697710 6461 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0214 11:10:57.697715 6461 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697733 6461 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0214 11:10:57.697811 6461 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2gwlw_openshift-ovn-kubernetes(1ca5acc0-8b15-4052-883f-1987b7ca0067)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.758856 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac91dd23-e4e5-4073-af76-d760dfdd1adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fz5f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.808696 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.808751 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.808764 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.808781 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.808793 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:59Z","lastTransitionTime":"2026-02-14T11:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.826202 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 21:41:33.985569201 +0000 UTC Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.835708 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:10:59 crc kubenswrapper[4904]: E0214 11:10:59.835816 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.836094 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.836160 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:10:59 crc kubenswrapper[4904]: E0214 11:10:59.836223 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:10:59 crc kubenswrapper[4904]: E0214 11:10:59.836300 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.848346 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.869228 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"message\\\":\\\"ift-kube-scheduler-crc openshift-multus/network-metrics-daemon-fz5f2 openshift-network-operator/iptables-alerter-4ln5h openshift-network-diagnostics/network-check-target-xd92c openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI0214 11:10:57.697651 6461 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI0214 11:10:57.697673 6461 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697684 6461 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697702 6461 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0214 11:10:57.697710 6461 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0214 11:10:57.697715 6461 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697733 6461 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0214 11:10:57.697811 6461 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2gwlw_openshift-ovn-kubernetes(1ca5acc0-8b15-4052-883f-1987b7ca0067)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.883706 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac91dd23-e4e5-4073-af76-d760dfdd1adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fz5f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.897085 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfcec17d-4d9a-4cdc-a86c-d05e21bedefb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee16aacae98d845ea68a860c26436a43d376f74ea7abd4396984fa63a836cb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc073ab93f3609414e270a023f8d2bad18f489eb6b37ac9a15cd6bf8b4dfbbf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2411332719f59b693409f1e2c9a10c280a102c11216e67d8eea8bfa8bdfe809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.910827 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.910907 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.910928 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.910951 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.910970 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:10:59Z","lastTransitionTime":"2026-02-14T11:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.911044 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.927980 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.947589 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.966521 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:10:59 crc kubenswrapper[4904]: I0214 11:10:59.986174 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:10:59Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.003136 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:00Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.013667 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.013716 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.013732 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.013756 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.013776 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:00Z","lastTransitionTime":"2026-02-14T11:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.020548 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f26819f5-4042-4943-b973-dba56706542e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7f9f319d3a100ef0c94c354ea3722a74d3541a399fe8ccedc4185d34934cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae046a4c4d0f2475bf13ed37f1f9c2029c5b73d436b31705e18ed030f7fc4b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f5pqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:00Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.034985 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:00Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.050688 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:00Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.071565 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:00Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.083492 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:00Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.104318 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:00Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.115777 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.115911 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.115923 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.115939 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.115948 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:00Z","lastTransitionTime":"2026-02-14T11:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.125183 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:00Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.217923 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.217973 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.217983 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.217998 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.218008 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:00Z","lastTransitionTime":"2026-02-14T11:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.320269 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.320322 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.320339 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.320362 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.320383 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:00Z","lastTransitionTime":"2026-02-14T11:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.422188 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.422230 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.422243 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.422263 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.422274 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:00Z","lastTransitionTime":"2026-02-14T11:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.524330 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.524370 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.524386 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.524406 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.524421 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:00Z","lastTransitionTime":"2026-02-14T11:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.623380 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs\") pod \"network-metrics-daemon-fz5f2\" (UID: \"ac91dd23-e4e5-4073-af76-d760dfdd1adc\") " pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:00 crc kubenswrapper[4904]: E0214 11:11:00.623510 4904 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 14 11:11:00 crc kubenswrapper[4904]: E0214 11:11:00.623587 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs podName:ac91dd23-e4e5-4073-af76-d760dfdd1adc nodeName:}" failed. No retries permitted until 2026-02-14 11:11:16.62356693 +0000 UTC m=+67.436331601 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs") pod "network-metrics-daemon-fz5f2" (UID: "ac91dd23-e4e5-4073-af76-d760dfdd1adc") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.627016 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.627043 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.627052 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.627065 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.627076 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:00Z","lastTransitionTime":"2026-02-14T11:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.729113 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.729150 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.729161 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.729177 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.729189 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:00Z","lastTransitionTime":"2026-02-14T11:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.827259 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 13:16:20.058062289 +0000 UTC Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.831277 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.831338 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.831360 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.831389 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.831409 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:00Z","lastTransitionTime":"2026-02-14T11:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.835300 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:00 crc kubenswrapper[4904]: E0214 11:11:00.835410 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.934488 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.934525 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.934539 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.934555 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:00 crc kubenswrapper[4904]: I0214 11:11:00.934566 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:00Z","lastTransitionTime":"2026-02-14T11:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.037052 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.037122 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.037147 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.037184 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.037206 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:01Z","lastTransitionTime":"2026-02-14T11:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.138913 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.138982 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.138994 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.139010 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.139023 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:01Z","lastTransitionTime":"2026-02-14T11:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.241624 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.241685 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.241698 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.241714 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.241725 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:01Z","lastTransitionTime":"2026-02-14T11:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.344811 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.344868 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.344881 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.344897 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.344936 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:01Z","lastTransitionTime":"2026-02-14T11:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.447612 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.447650 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.447659 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.447675 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.447685 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:01Z","lastTransitionTime":"2026-02-14T11:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.550444 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.550573 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.554850 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.554870 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.554882 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:01Z","lastTransitionTime":"2026-02-14T11:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.657115 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.657161 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.657178 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.657200 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.657219 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:01Z","lastTransitionTime":"2026-02-14T11:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.737043 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:11:01 crc kubenswrapper[4904]: E0214 11:11:01.737217 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:11:33.737197773 +0000 UTC m=+84.549962424 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.759248 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.759283 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.759298 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.759318 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.759333 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:01Z","lastTransitionTime":"2026-02-14T11:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.828189 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 13:22:33.082925649 +0000 UTC Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.835538 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:01 crc kubenswrapper[4904]: E0214 11:11:01.835674 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.835578 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:01 crc kubenswrapper[4904]: E0214 11:11:01.835753 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.835538 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:01 crc kubenswrapper[4904]: E0214 11:11:01.835812 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.838332 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.838377 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.838398 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.838429 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:01 crc kubenswrapper[4904]: E0214 11:11:01.838501 4904 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 14 11:11:01 crc kubenswrapper[4904]: E0214 11:11:01.838543 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-14 11:11:33.838530455 +0000 UTC m=+84.651295116 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 14 11:11:01 crc kubenswrapper[4904]: E0214 11:11:01.838604 4904 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 14 11:11:01 crc kubenswrapper[4904]: E0214 11:11:01.838631 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-14 11:11:33.838622687 +0000 UTC m=+84.651387358 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 14 11:11:01 crc kubenswrapper[4904]: E0214 11:11:01.838693 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 14 11:11:01 crc kubenswrapper[4904]: E0214 11:11:01.838707 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 14 11:11:01 crc kubenswrapper[4904]: E0214 11:11:01.838718 4904 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:11:01 crc kubenswrapper[4904]: E0214 11:11:01.838745 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-14 11:11:33.838736579 +0000 UTC m=+84.651501240 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:11:01 crc kubenswrapper[4904]: E0214 11:11:01.838798 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 14 11:11:01 crc kubenswrapper[4904]: E0214 11:11:01.838812 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 14 11:11:01 crc kubenswrapper[4904]: E0214 11:11:01.838820 4904 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:11:01 crc kubenswrapper[4904]: E0214 11:11:01.838862 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-14 11:11:33.838853972 +0000 UTC m=+84.651618643 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.860967 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.860988 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.860996 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.861007 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.861016 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:01Z","lastTransitionTime":"2026-02-14T11:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.962721 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.962753 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.962763 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.962778 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:01 crc kubenswrapper[4904]: I0214 11:11:01.962788 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:01Z","lastTransitionTime":"2026-02-14T11:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.064865 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.064903 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.064912 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.064927 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.064936 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:02Z","lastTransitionTime":"2026-02-14T11:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.167549 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.167578 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.167589 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.167603 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.167615 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:02Z","lastTransitionTime":"2026-02-14T11:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.269654 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.269712 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.269726 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.269744 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.269755 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:02Z","lastTransitionTime":"2026-02-14T11:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.372911 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.372970 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.372992 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.373020 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.373042 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:02Z","lastTransitionTime":"2026-02-14T11:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.475305 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.475336 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.475347 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.475359 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.475370 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:02Z","lastTransitionTime":"2026-02-14T11:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.577707 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.577758 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.577774 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.577797 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.577812 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:02Z","lastTransitionTime":"2026-02-14T11:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.679584 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.679638 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.679651 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.679670 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.679683 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:02Z","lastTransitionTime":"2026-02-14T11:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.781785 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.781810 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.781820 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.781858 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.781870 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:02Z","lastTransitionTime":"2026-02-14T11:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.828733 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 18:15:47.932702405 +0000 UTC Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.836259 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:02 crc kubenswrapper[4904]: E0214 11:11:02.836390 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.884383 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.884421 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.884429 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.884442 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.884451 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:02Z","lastTransitionTime":"2026-02-14T11:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.986693 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.986729 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.986740 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.986758 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:02 crc kubenswrapper[4904]: I0214 11:11:02.986770 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:02Z","lastTransitionTime":"2026-02-14T11:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.089042 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.089086 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.089107 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.089129 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.089144 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:03Z","lastTransitionTime":"2026-02-14T11:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.191748 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.191789 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.191799 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.191814 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.191826 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:03Z","lastTransitionTime":"2026-02-14T11:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.293522 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.293557 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.293571 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.293587 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.293598 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:03Z","lastTransitionTime":"2026-02-14T11:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.395652 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.395982 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.396107 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.396282 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.396411 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:03Z","lastTransitionTime":"2026-02-14T11:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.498131 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.498417 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.498426 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.498438 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.498446 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:03Z","lastTransitionTime":"2026-02-14T11:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.601481 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.601740 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.601810 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.601933 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.602000 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:03Z","lastTransitionTime":"2026-02-14T11:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.704187 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.704716 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.704782 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.704870 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.704961 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:03Z","lastTransitionTime":"2026-02-14T11:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.807238 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.807278 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.807288 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.807304 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.807316 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:03Z","lastTransitionTime":"2026-02-14T11:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.829526 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 21:27:05.777591641 +0000 UTC Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.835859 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.835882 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.835899 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:03 crc kubenswrapper[4904]: E0214 11:11:03.836365 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:03 crc kubenswrapper[4904]: E0214 11:11:03.836458 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:03 crc kubenswrapper[4904]: E0214 11:11:03.836524 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.910261 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.910301 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.910309 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.910323 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:03 crc kubenswrapper[4904]: I0214 11:11:03.910333 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:03Z","lastTransitionTime":"2026-02-14T11:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.013163 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.013513 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.013693 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.013940 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.014116 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:04Z","lastTransitionTime":"2026-02-14T11:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.116754 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.116793 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.116804 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.116819 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.116828 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:04Z","lastTransitionTime":"2026-02-14T11:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.219468 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.219816 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.220039 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.220181 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.220353 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:04Z","lastTransitionTime":"2026-02-14T11:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.323753 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.323823 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.323892 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.323923 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.323947 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:04Z","lastTransitionTime":"2026-02-14T11:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.427238 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.427478 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.427607 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.427711 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.427792 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:04Z","lastTransitionTime":"2026-02-14T11:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.530128 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.530434 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.530519 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.530656 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.530781 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:04Z","lastTransitionTime":"2026-02-14T11:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.633032 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.633102 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.633129 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.633160 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.633184 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:04Z","lastTransitionTime":"2026-02-14T11:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.736088 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.736928 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.737140 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.737352 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.737552 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:04Z","lastTransitionTime":"2026-02-14T11:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.829928 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 12:06:44.407647449 +0000 UTC Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.836329 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:04 crc kubenswrapper[4904]: E0214 11:11:04.836536 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.840331 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.840370 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.840382 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.840401 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.840415 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:04Z","lastTransitionTime":"2026-02-14T11:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.942625 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.942944 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.943065 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.943170 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:04 crc kubenswrapper[4904]: I0214 11:11:04.943305 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:04Z","lastTransitionTime":"2026-02-14T11:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.045860 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.045916 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.045928 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.045945 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.045956 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:05Z","lastTransitionTime":"2026-02-14T11:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.147654 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.147923 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.148024 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.148126 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.148267 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:05Z","lastTransitionTime":"2026-02-14T11:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.251452 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.251791 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.251936 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.252090 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.252215 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:05Z","lastTransitionTime":"2026-02-14T11:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.354683 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.354734 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.354744 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.354757 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.354767 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:05Z","lastTransitionTime":"2026-02-14T11:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.457644 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.457917 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.458073 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.458170 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.458245 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:05Z","lastTransitionTime":"2026-02-14T11:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.561369 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.561439 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.561463 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.561487 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.561504 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:05Z","lastTransitionTime":"2026-02-14T11:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.663962 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.664261 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.664472 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.664643 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.664797 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:05Z","lastTransitionTime":"2026-02-14T11:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.767368 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.767438 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.767461 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.767489 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.767510 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:05Z","lastTransitionTime":"2026-02-14T11:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.830182 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 13:32:26.139302559 +0000 UTC Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.835909 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.835973 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:05 crc kubenswrapper[4904]: E0214 11:11:05.836070 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:05 crc kubenswrapper[4904]: E0214 11:11:05.836179 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.836427 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:05 crc kubenswrapper[4904]: E0214 11:11:05.836696 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.869869 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.869928 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.869945 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.869967 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.869984 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:05Z","lastTransitionTime":"2026-02-14T11:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.972602 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.972669 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.972709 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.972737 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:05 crc kubenswrapper[4904]: I0214 11:11:05.972757 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:05Z","lastTransitionTime":"2026-02-14T11:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.075260 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.075307 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.075324 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.075344 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.075362 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:06Z","lastTransitionTime":"2026-02-14T11:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.124596 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.124663 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.124679 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.124702 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.124717 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:06Z","lastTransitionTime":"2026-02-14T11:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:06 crc kubenswrapper[4904]: E0214 11:11:06.141432 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:06Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.151226 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.151665 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.151686 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.151709 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.151726 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:06Z","lastTransitionTime":"2026-02-14T11:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:06 crc kubenswrapper[4904]: E0214 11:11:06.170167 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:06Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.174422 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.174453 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.174462 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.174475 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.174485 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:06Z","lastTransitionTime":"2026-02-14T11:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:06 crc kubenswrapper[4904]: E0214 11:11:06.189052 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:06Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.193634 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.193673 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.193683 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.193696 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.193707 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:06Z","lastTransitionTime":"2026-02-14T11:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:06 crc kubenswrapper[4904]: E0214 11:11:06.215298 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:06Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.220335 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.220373 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.220386 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.220404 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.220418 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:06Z","lastTransitionTime":"2026-02-14T11:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:06 crc kubenswrapper[4904]: E0214 11:11:06.233319 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:06Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:06 crc kubenswrapper[4904]: E0214 11:11:06.234018 4904 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.236452 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.236485 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.236498 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.236515 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.236526 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:06Z","lastTransitionTime":"2026-02-14T11:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.340321 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.340375 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.340390 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.340410 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.340423 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:06Z","lastTransitionTime":"2026-02-14T11:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.442770 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.442806 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.442814 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.442827 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.442867 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:06Z","lastTransitionTime":"2026-02-14T11:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.545335 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.545425 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.545436 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.545451 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.545460 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:06Z","lastTransitionTime":"2026-02-14T11:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.648173 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.648235 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.648247 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.648268 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.648298 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:06Z","lastTransitionTime":"2026-02-14T11:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.750742 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.750777 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.750787 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.750801 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.750811 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:06Z","lastTransitionTime":"2026-02-14T11:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.830799 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 12:52:02.763775482 +0000 UTC Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.836176 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:06 crc kubenswrapper[4904]: E0214 11:11:06.836333 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.854697 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.854754 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.854766 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.854781 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.854793 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:06Z","lastTransitionTime":"2026-02-14T11:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.957735 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.957779 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.957788 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.957802 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:06 crc kubenswrapper[4904]: I0214 11:11:06.957812 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:06Z","lastTransitionTime":"2026-02-14T11:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.059795 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.059881 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.059899 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.059924 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.059939 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:07Z","lastTransitionTime":"2026-02-14T11:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.162133 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.162210 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.162232 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.162258 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.162275 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:07Z","lastTransitionTime":"2026-02-14T11:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.264926 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.265000 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.265027 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.265059 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.265086 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:07Z","lastTransitionTime":"2026-02-14T11:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.367430 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.367478 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.367497 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.367518 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.367535 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:07Z","lastTransitionTime":"2026-02-14T11:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.470560 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.470607 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.470623 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.470647 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.470663 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:07Z","lastTransitionTime":"2026-02-14T11:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.572619 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.572687 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.572698 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.572713 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.572723 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:07Z","lastTransitionTime":"2026-02-14T11:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.674693 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.674738 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.674751 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.674772 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.674785 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:07Z","lastTransitionTime":"2026-02-14T11:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.777637 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.777681 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.777699 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.777722 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.777737 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:07Z","lastTransitionTime":"2026-02-14T11:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.830989 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 02:31:21.044216727 +0000 UTC Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.836269 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.836318 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.836402 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:07 crc kubenswrapper[4904]: E0214 11:11:07.836534 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:07 crc kubenswrapper[4904]: E0214 11:11:07.836823 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:07 crc kubenswrapper[4904]: E0214 11:11:07.837152 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.880378 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.880408 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.880417 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.880430 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.880440 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:07Z","lastTransitionTime":"2026-02-14T11:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.982059 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.982096 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.982107 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.982123 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:07 crc kubenswrapper[4904]: I0214 11:11:07.982133 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:07Z","lastTransitionTime":"2026-02-14T11:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.084172 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.084204 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.084212 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.084246 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.084258 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:08Z","lastTransitionTime":"2026-02-14T11:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.186853 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.186960 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.186975 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.187008 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.187021 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:08Z","lastTransitionTime":"2026-02-14T11:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.290395 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.290460 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.290478 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.290509 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.290531 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:08Z","lastTransitionTime":"2026-02-14T11:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.393380 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.393421 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.393429 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.393444 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.393453 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:08Z","lastTransitionTime":"2026-02-14T11:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.495354 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.495386 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.495397 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.495413 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.495424 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:08Z","lastTransitionTime":"2026-02-14T11:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.597570 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.597602 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.597613 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.597628 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.597638 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:08Z","lastTransitionTime":"2026-02-14T11:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.700082 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.700138 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.700150 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.700165 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.700175 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:08Z","lastTransitionTime":"2026-02-14T11:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.801967 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.802036 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.802054 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.802068 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.802080 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:08Z","lastTransitionTime":"2026-02-14T11:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.831436 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 12:34:46.748953138 +0000 UTC Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.835709 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:08 crc kubenswrapper[4904]: E0214 11:11:08.835828 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.903903 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.903953 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.903963 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.903976 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:08 crc kubenswrapper[4904]: I0214 11:11:08.903985 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:08Z","lastTransitionTime":"2026-02-14T11:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.006159 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.006190 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.006199 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.006216 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.006226 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:09Z","lastTransitionTime":"2026-02-14T11:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.108861 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.108903 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.108918 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.108937 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.108948 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:09Z","lastTransitionTime":"2026-02-14T11:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.211218 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.211280 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.211291 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.211308 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.211318 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:09Z","lastTransitionTime":"2026-02-14T11:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.313684 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.313724 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.313736 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.313750 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.313759 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:09Z","lastTransitionTime":"2026-02-14T11:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.415939 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.415981 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.415998 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.416015 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.416025 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:09Z","lastTransitionTime":"2026-02-14T11:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.518373 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.518426 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.518438 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.518455 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.518469 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:09Z","lastTransitionTime":"2026-02-14T11:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.620712 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.620766 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.620782 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.620802 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.620820 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:09Z","lastTransitionTime":"2026-02-14T11:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.723356 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.723400 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.723412 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.723428 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.723439 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:09Z","lastTransitionTime":"2026-02-14T11:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.825297 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.825326 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.825335 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.825348 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.825356 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:09Z","lastTransitionTime":"2026-02-14T11:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.831544 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 14:35:46.877559973 +0000 UTC Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.835883 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.836130 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:09 crc kubenswrapper[4904]: E0214 11:11:09.836118 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.836141 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:09 crc kubenswrapper[4904]: E0214 11:11:09.836535 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:09 crc kubenswrapper[4904]: E0214 11:11:09.836334 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.850054 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:09Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.862152 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f26819f5-4042-4943-b973-dba56706542e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7f9f319d3a100ef0c94c354ea3722a74d3541a399fe8ccedc4185d34934cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae046a4c4d0f2475bf13ed37f1f9c2029c5b73d436b31705e18ed030f7fc4b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f5pqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:09Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.873796 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:09Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.885083 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:09Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.895291 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:09Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.905090 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:09Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.917797 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:09Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.926985 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.927018 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.927045 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.927060 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.927070 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:09Z","lastTransitionTime":"2026-02-14T11:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.927239 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:09Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.935051 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:09Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.951275 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"message\\\":\\\"ift-kube-scheduler-crc openshift-multus/network-metrics-daemon-fz5f2 openshift-network-operator/iptables-alerter-4ln5h openshift-network-diagnostics/network-check-target-xd92c openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI0214 11:10:57.697651 6461 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI0214 11:10:57.697673 6461 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697684 6461 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697702 6461 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0214 11:10:57.697710 6461 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0214 11:10:57.697715 6461 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697733 6461 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0214 11:10:57.697811 6461 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2gwlw_openshift-ovn-kubernetes(1ca5acc0-8b15-4052-883f-1987b7ca0067)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:09Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.960386 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac91dd23-e4e5-4073-af76-d760dfdd1adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fz5f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:09Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.971630 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:09Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.982786 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:09Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:09 crc kubenswrapper[4904]: I0214 11:11:09.993215 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfcec17d-4d9a-4cdc-a86c-d05e21bedefb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee16aacae98d845ea68a860c26436a43d376f74ea7abd4396984fa63a836cb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc073ab93f3609414e270a023f8d2bad18f489eb6b37ac9a15cd6bf8b4dfbbf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2411332719f59b693409f1e2c9a10c280a102c11216e67d8eea8bfa8bdfe809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:09Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.003935 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:10Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.023864 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:10Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.028845 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.028878 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.028886 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.028900 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.028924 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:10Z","lastTransitionTime":"2026-02-14T11:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.037753 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:10Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.132870 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.132906 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.132915 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.132931 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.132942 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:10Z","lastTransitionTime":"2026-02-14T11:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.235504 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.235555 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.235566 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.235584 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.235595 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:10Z","lastTransitionTime":"2026-02-14T11:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.337243 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.337272 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.337282 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.337295 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.337306 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:10Z","lastTransitionTime":"2026-02-14T11:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.439226 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.439254 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.439264 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.439276 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.439285 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:10Z","lastTransitionTime":"2026-02-14T11:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.541627 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.541669 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.541678 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.541692 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.541702 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:10Z","lastTransitionTime":"2026-02-14T11:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.644153 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.644181 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.644189 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.644201 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.644211 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:10Z","lastTransitionTime":"2026-02-14T11:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.746269 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.746301 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.746309 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.746321 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.746330 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:10Z","lastTransitionTime":"2026-02-14T11:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.831893 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 01:11:58.010581363 +0000 UTC Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.836256 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:10 crc kubenswrapper[4904]: E0214 11:11:10.836374 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.848682 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.848711 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.848720 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.848734 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.848743 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:10Z","lastTransitionTime":"2026-02-14T11:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.981637 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.981665 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.981673 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.981684 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:10 crc kubenswrapper[4904]: I0214 11:11:10.981692 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:10Z","lastTransitionTime":"2026-02-14T11:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.084186 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.084216 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.084224 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.084237 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.084245 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:11Z","lastTransitionTime":"2026-02-14T11:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.186052 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.186108 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.186121 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.186155 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.186166 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:11Z","lastTransitionTime":"2026-02-14T11:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.289872 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.289913 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.289925 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.289943 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.289956 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:11Z","lastTransitionTime":"2026-02-14T11:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.392641 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.392694 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.392706 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.392722 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.392732 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:11Z","lastTransitionTime":"2026-02-14T11:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.494680 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.494764 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.494782 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.494811 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.494831 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:11Z","lastTransitionTime":"2026-02-14T11:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.598012 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.598045 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.598055 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.598071 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.598080 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:11Z","lastTransitionTime":"2026-02-14T11:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.700971 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.701014 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.701024 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.701040 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.701051 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:11Z","lastTransitionTime":"2026-02-14T11:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.803480 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.803516 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.803524 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.803539 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.803550 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:11Z","lastTransitionTime":"2026-02-14T11:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.832817 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 15:43:52.533722762 +0000 UTC Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.837051 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.837110 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:11 crc kubenswrapper[4904]: E0214 11:11:11.837173 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.837188 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:11 crc kubenswrapper[4904]: E0214 11:11:11.837406 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:11 crc kubenswrapper[4904]: E0214 11:11:11.837498 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.905942 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.905989 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.906002 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.906018 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:11 crc kubenswrapper[4904]: I0214 11:11:11.906030 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:11Z","lastTransitionTime":"2026-02-14T11:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.010801 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.010869 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.010883 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.010899 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.010909 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:12Z","lastTransitionTime":"2026-02-14T11:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.112922 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.112956 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.112965 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.112979 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.112989 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:12Z","lastTransitionTime":"2026-02-14T11:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.215241 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.215280 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.215289 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.215302 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.215312 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:12Z","lastTransitionTime":"2026-02-14T11:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.317293 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.317349 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.317359 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.317374 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.317384 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:12Z","lastTransitionTime":"2026-02-14T11:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.419398 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.419430 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.419441 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.419456 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.419466 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:12Z","lastTransitionTime":"2026-02-14T11:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.521399 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.521433 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.521442 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.521456 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.521464 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:12Z","lastTransitionTime":"2026-02-14T11:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.623997 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.624058 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.624069 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.624084 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.624095 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:12Z","lastTransitionTime":"2026-02-14T11:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.726477 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.726520 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.726533 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.726549 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.726562 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:12Z","lastTransitionTime":"2026-02-14T11:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.828412 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.828451 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.828463 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.828480 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.828493 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:12Z","lastTransitionTime":"2026-02-14T11:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.833571 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 01:55:43.910637596 +0000 UTC Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.835919 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:12 crc kubenswrapper[4904]: E0214 11:11:12.836259 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.836542 4904 scope.go:117] "RemoveContainer" containerID="f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8" Feb 14 11:11:12 crc kubenswrapper[4904]: E0214 11:11:12.836749 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2gwlw_openshift-ovn-kubernetes(1ca5acc0-8b15-4052-883f-1987b7ca0067)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.930888 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.930941 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.930955 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.930975 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:12 crc kubenswrapper[4904]: I0214 11:11:12.930988 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:12Z","lastTransitionTime":"2026-02-14T11:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.033428 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.033478 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.033489 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.033506 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.033518 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:13Z","lastTransitionTime":"2026-02-14T11:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.135604 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.135638 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.135650 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.135665 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.135675 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:13Z","lastTransitionTime":"2026-02-14T11:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.237591 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.237627 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.237637 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.237651 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.237660 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:13Z","lastTransitionTime":"2026-02-14T11:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.340034 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.340070 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.340081 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.340097 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.340108 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:13Z","lastTransitionTime":"2026-02-14T11:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.442291 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.442328 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.442340 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.442354 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.442366 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:13Z","lastTransitionTime":"2026-02-14T11:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.544762 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.544802 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.544812 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.544828 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.544855 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:13Z","lastTransitionTime":"2026-02-14T11:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.646766 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.646800 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.646808 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.646825 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.646850 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:13Z","lastTransitionTime":"2026-02-14T11:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.748726 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.748760 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.748771 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.748786 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.748797 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:13Z","lastTransitionTime":"2026-02-14T11:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.834523 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 08:58:29.384671047 +0000 UTC Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.835866 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.835884 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.835956 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:13 crc kubenswrapper[4904]: E0214 11:11:13.836118 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:13 crc kubenswrapper[4904]: E0214 11:11:13.836210 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:13 crc kubenswrapper[4904]: E0214 11:11:13.836275 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.851120 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.851162 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.851174 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.851191 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.851204 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:13Z","lastTransitionTime":"2026-02-14T11:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.953436 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.953465 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.953475 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.953487 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:13 crc kubenswrapper[4904]: I0214 11:11:13.953496 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:13Z","lastTransitionTime":"2026-02-14T11:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.056398 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.056428 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.056436 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.056453 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.056464 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:14Z","lastTransitionTime":"2026-02-14T11:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.158620 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.158658 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.158669 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.158684 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.158693 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:14Z","lastTransitionTime":"2026-02-14T11:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.260520 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.260559 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.260571 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.260587 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.260600 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:14Z","lastTransitionTime":"2026-02-14T11:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.362784 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.362822 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.362848 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.362864 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.362877 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:14Z","lastTransitionTime":"2026-02-14T11:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.465520 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.465555 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.465564 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.465579 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.465589 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:14Z","lastTransitionTime":"2026-02-14T11:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.568787 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.568855 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.568868 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.568884 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.568896 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:14Z","lastTransitionTime":"2026-02-14T11:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.671192 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.671246 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.671256 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.671270 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.671280 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:14Z","lastTransitionTime":"2026-02-14T11:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.773061 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.773102 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.773112 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.773127 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.773139 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:14Z","lastTransitionTime":"2026-02-14T11:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.835471 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 18:37:19.818510084 +0000 UTC Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.835506 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:14 crc kubenswrapper[4904]: E0214 11:11:14.835645 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.875212 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.875260 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.875270 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.875285 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.875295 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:14Z","lastTransitionTime":"2026-02-14T11:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.977234 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.977262 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.977286 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.977300 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:14 crc kubenswrapper[4904]: I0214 11:11:14.977308 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:14Z","lastTransitionTime":"2026-02-14T11:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.079996 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.080032 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.080043 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.080059 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.080070 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:15Z","lastTransitionTime":"2026-02-14T11:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.181787 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.181817 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.181826 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.181852 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.181865 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:15Z","lastTransitionTime":"2026-02-14T11:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.284761 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.284795 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.284805 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.284819 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.284846 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:15Z","lastTransitionTime":"2026-02-14T11:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.387781 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.387824 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.387848 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.387863 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.387876 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:15Z","lastTransitionTime":"2026-02-14T11:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.489584 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.489792 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.489882 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.489980 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.490077 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:15Z","lastTransitionTime":"2026-02-14T11:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.592130 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.592159 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.592168 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.592186 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.592198 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:15Z","lastTransitionTime":"2026-02-14T11:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.694387 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.694415 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.694423 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.694436 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.694444 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:15Z","lastTransitionTime":"2026-02-14T11:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.796880 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.796916 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.796927 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.796942 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.796953 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:15Z","lastTransitionTime":"2026-02-14T11:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.835476 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.835510 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.835527 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.835576 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 07:56:58.289665862 +0000 UTC Feb 14 11:11:15 crc kubenswrapper[4904]: E0214 11:11:15.835636 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:15 crc kubenswrapper[4904]: E0214 11:11:15.835746 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:15 crc kubenswrapper[4904]: E0214 11:11:15.835853 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.898732 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.898764 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.898771 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.898785 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:15 crc kubenswrapper[4904]: I0214 11:11:15.898794 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:15Z","lastTransitionTime":"2026-02-14T11:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.000523 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.000562 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.000572 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.000585 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.000594 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:16Z","lastTransitionTime":"2026-02-14T11:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.103363 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.103404 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.103413 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.103427 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.103436 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:16Z","lastTransitionTime":"2026-02-14T11:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.204971 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.205000 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.205010 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.205022 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.205032 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:16Z","lastTransitionTime":"2026-02-14T11:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.306940 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.306967 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.306975 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.306987 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.306997 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:16Z","lastTransitionTime":"2026-02-14T11:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.408123 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.408147 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.408155 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.408168 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.408176 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:16Z","lastTransitionTime":"2026-02-14T11:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.419796 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.419852 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.419863 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.419880 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.419892 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:16Z","lastTransitionTime":"2026-02-14T11:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:16 crc kubenswrapper[4904]: E0214 11:11:16.430944 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:16Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.433803 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.433856 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.433867 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.433886 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.433897 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:16Z","lastTransitionTime":"2026-02-14T11:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:16 crc kubenswrapper[4904]: E0214 11:11:16.448206 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:16Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.451065 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.451103 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.451112 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.451125 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.451136 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:16Z","lastTransitionTime":"2026-02-14T11:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:16 crc kubenswrapper[4904]: E0214 11:11:16.461559 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:16Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.464399 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.464425 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.464435 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.464448 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.464457 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:16Z","lastTransitionTime":"2026-02-14T11:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:16 crc kubenswrapper[4904]: E0214 11:11:16.476688 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:16Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.479928 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.480044 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.480107 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.480195 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.480256 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:16Z","lastTransitionTime":"2026-02-14T11:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:16 crc kubenswrapper[4904]: E0214 11:11:16.491530 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:16Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:16 crc kubenswrapper[4904]: E0214 11:11:16.491640 4904 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.509612 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.509640 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.509650 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.509664 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.509674 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:16Z","lastTransitionTime":"2026-02-14T11:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.611616 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.611899 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.611990 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.612126 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.612214 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:16Z","lastTransitionTime":"2026-02-14T11:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.686407 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs\") pod \"network-metrics-daemon-fz5f2\" (UID: \"ac91dd23-e4e5-4073-af76-d760dfdd1adc\") " pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:16 crc kubenswrapper[4904]: E0214 11:11:16.686575 4904 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 14 11:11:16 crc kubenswrapper[4904]: E0214 11:11:16.686660 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs podName:ac91dd23-e4e5-4073-af76-d760dfdd1adc nodeName:}" failed. No retries permitted until 2026-02-14 11:11:48.686643007 +0000 UTC m=+99.499407668 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs") pod "network-metrics-daemon-fz5f2" (UID: "ac91dd23-e4e5-4073-af76-d760dfdd1adc") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.714199 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.714238 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.714249 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.714267 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.714279 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:16Z","lastTransitionTime":"2026-02-14T11:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.816872 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.816923 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.816934 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.816955 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.816967 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:16Z","lastTransitionTime":"2026-02-14T11:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.835347 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:16 crc kubenswrapper[4904]: E0214 11:11:16.835467 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.835785 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 14:29:20.884128035 +0000 UTC Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.918392 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.918416 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.918424 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.918436 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:16 crc kubenswrapper[4904]: I0214 11:11:16.918444 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:16Z","lastTransitionTime":"2026-02-14T11:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.020006 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.020036 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.020045 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.020075 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.020087 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:17Z","lastTransitionTime":"2026-02-14T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.122714 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.122930 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.122997 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.123069 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.123125 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:17Z","lastTransitionTime":"2026-02-14T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.225158 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.225382 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.225446 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.225508 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.225572 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:17Z","lastTransitionTime":"2026-02-14T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.327638 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.327880 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.327976 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.328073 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.328155 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:17Z","lastTransitionTime":"2026-02-14T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.430036 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.430064 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.430073 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.430085 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.430095 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:17Z","lastTransitionTime":"2026-02-14T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.532242 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.532275 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.532286 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.532301 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.532313 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:17Z","lastTransitionTime":"2026-02-14T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.633764 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.633800 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.633810 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.633826 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.633857 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:17Z","lastTransitionTime":"2026-02-14T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.735470 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.735507 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.735519 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.735533 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.735543 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:17Z","lastTransitionTime":"2026-02-14T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.835304 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:17 crc kubenswrapper[4904]: E0214 11:11:17.835399 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.835437 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.835456 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:17 crc kubenswrapper[4904]: E0214 11:11:17.835497 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:17 crc kubenswrapper[4904]: E0214 11:11:17.835755 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.835974 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 09:32:26.488679401 +0000 UTC Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.836969 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.836985 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.836994 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.837005 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.837015 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:17Z","lastTransitionTime":"2026-02-14T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.939288 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.939324 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.939332 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.939347 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:17 crc kubenswrapper[4904]: I0214 11:11:17.939355 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:17Z","lastTransitionTime":"2026-02-14T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.041470 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.041504 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.041512 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.041524 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.041533 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:18Z","lastTransitionTime":"2026-02-14T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.143594 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.143627 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.143635 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.143647 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.143656 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:18Z","lastTransitionTime":"2026-02-14T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.246321 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.246353 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.246362 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.246379 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.246388 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:18Z","lastTransitionTime":"2026-02-14T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.349147 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.349198 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.349224 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.349246 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.349260 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:18Z","lastTransitionTime":"2026-02-14T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.451372 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.451428 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.451440 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.451459 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.451473 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:18Z","lastTransitionTime":"2026-02-14T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.554716 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.554746 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.554764 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.554778 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.554788 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:18Z","lastTransitionTime":"2026-02-14T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.657611 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.658172 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.658370 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.658661 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.658903 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:18Z","lastTransitionTime":"2026-02-14T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.762711 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.762794 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.762807 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.762826 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.762859 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:18Z","lastTransitionTime":"2026-02-14T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.835780 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:18 crc kubenswrapper[4904]: E0214 11:11:18.835951 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.836068 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 15:31:36.790984595 +0000 UTC Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.865273 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.865312 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.865323 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.865344 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.865355 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:18Z","lastTransitionTime":"2026-02-14T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.968648 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.968690 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.968699 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.968714 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:18 crc kubenswrapper[4904]: I0214 11:11:18.968723 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:18Z","lastTransitionTime":"2026-02-14T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.071796 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.071870 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.071882 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.071896 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.071908 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:19Z","lastTransitionTime":"2026-02-14T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.174184 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.174247 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.174257 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.174269 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.174279 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:19Z","lastTransitionTime":"2026-02-14T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.277095 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.277184 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.277199 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.277217 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.277235 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:19Z","lastTransitionTime":"2026-02-14T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.381002 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.381286 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.381368 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.381451 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.381527 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:19Z","lastTransitionTime":"2026-02-14T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.484049 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.484085 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.484094 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.484108 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.484118 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:19Z","lastTransitionTime":"2026-02-14T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.517587 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jz5f5_0b01cc02-2e62-46e1-b07d-b3c0bff2b050/kube-multus/0.log" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.517633 4904 generic.go:334] "Generic (PLEG): container finished" podID="0b01cc02-2e62-46e1-b07d-b3c0bff2b050" containerID="9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938" exitCode=1 Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.517661 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jz5f5" event={"ID":"0b01cc02-2e62-46e1-b07d-b3c0bff2b050","Type":"ContainerDied","Data":"9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938"} Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.517993 4904 scope.go:117] "RemoveContainer" containerID="9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.529927 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.547493 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f26819f5-4042-4943-b973-dba56706542e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7f9f319d3a100ef0c94c354ea3722a74d3541a399fe8ccedc4185d34934cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae046a4c4d0f2475bf13ed37f1f9c2029c5b73d436b31705e18ed030f7fc4b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f5pqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.557058 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.569164 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.584458 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.586433 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.586525 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.586587 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.586653 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.586724 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:19Z","lastTransitionTime":"2026-02-14T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.596626 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.607901 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.622097 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:11:19Z\\\",\\\"message\\\":\\\"2026-02-14T11:10:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ae7e0868-7f5e-443d-b025-c96c2b00ebce\\\\n2026-02-14T11:10:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ae7e0868-7f5e-443d-b025-c96c2b00ebce to /host/opt/cni/bin/\\\\n2026-02-14T11:10:33Z [verbose] multus-daemon started\\\\n2026-02-14T11:10:33Z [verbose] Readiness Indicator file check\\\\n2026-02-14T11:11:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.631571 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.649465 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"message\\\":\\\"ift-kube-scheduler-crc openshift-multus/network-metrics-daemon-fz5f2 openshift-network-operator/iptables-alerter-4ln5h openshift-network-diagnostics/network-check-target-xd92c openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI0214 11:10:57.697651 6461 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI0214 11:10:57.697673 6461 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697684 6461 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697702 6461 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0214 11:10:57.697710 6461 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0214 11:10:57.697715 6461 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697733 6461 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0214 11:10:57.697811 6461 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2gwlw_openshift-ovn-kubernetes(1ca5acc0-8b15-4052-883f-1987b7ca0067)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.660124 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac91dd23-e4e5-4073-af76-d760dfdd1adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fz5f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.674256 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.685919 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.688369 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.688486 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.688555 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.688625 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.688692 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:19Z","lastTransitionTime":"2026-02-14T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.698507 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.710571 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfcec17d-4d9a-4cdc-a86c-d05e21bedefb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee16aacae98d845ea68a860c26436a43d376f74ea7abd4396984fa63a836cb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc073ab93f3609414e270a023f8d2bad18f489eb6b37ac9a15cd6bf8b4dfbbf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2411332719f59b693409f1e2c9a10c280a102c11216e67d8eea8bfa8bdfe809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.721667 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.733891 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.791435 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.791504 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.791517 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.791536 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.791549 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:19Z","lastTransitionTime":"2026-02-14T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.835780 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.835792 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:19 crc kubenswrapper[4904]: E0214 11:11:19.835917 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.836010 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.836286 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 07:00:57.66099243 +0000 UTC Feb 14 11:11:19 crc kubenswrapper[4904]: E0214 11:11:19.836054 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:19 crc kubenswrapper[4904]: E0214 11:11:19.836478 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.853619 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.873911 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.886804 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfcec17d-4d9a-4cdc-a86c-d05e21bedefb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee16aacae98d845ea68a860c26436a43d376f74ea7abd4396984fa63a836cb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc073ab93f3609414e270a023f8d2bad18f489eb6b37ac9a15cd6bf8b4dfbbf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2411332719f59b693409f1e2c9a10c280a102c11216e67d8eea8bfa8bdfe809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.893253 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.893301 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.893311 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.893324 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.893332 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:19Z","lastTransitionTime":"2026-02-14T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.905486 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.919247 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.940935 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.951139 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.963876 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f26819f5-4042-4943-b973-dba56706542e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7f9f319d3a100ef0c94c354ea3722a74d3541a399fe8ccedc4185d34934cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae046a4c4d0f2475bf13ed37f1f9c2029c5b73d436b31705e18ed030f7fc4b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f5pqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.975192 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.987058 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.994792 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.994975 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.995174 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.995320 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:19 crc kubenswrapper[4904]: I0214 11:11:19.995488 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:19Z","lastTransitionTime":"2026-02-14T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.000372 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:19Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.018094 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.029287 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:11:19Z\\\",\\\"message\\\":\\\"2026-02-14T11:10:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ae7e0868-7f5e-443d-b025-c96c2b00ebce\\\\n2026-02-14T11:10:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ae7e0868-7f5e-443d-b025-c96c2b00ebce to /host/opt/cni/bin/\\\\n2026-02-14T11:10:33Z [verbose] multus-daemon started\\\\n2026-02-14T11:10:33Z [verbose] Readiness Indicator file check\\\\n2026-02-14T11:11:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.038855 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.048003 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.065446 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"message\\\":\\\"ift-kube-scheduler-crc openshift-multus/network-metrics-daemon-fz5f2 openshift-network-operator/iptables-alerter-4ln5h openshift-network-diagnostics/network-check-target-xd92c openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI0214 11:10:57.697651 6461 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI0214 11:10:57.697673 6461 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697684 6461 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697702 6461 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0214 11:10:57.697710 6461 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0214 11:10:57.697715 6461 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697733 6461 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0214 11:10:57.697811 6461 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2gwlw_openshift-ovn-kubernetes(1ca5acc0-8b15-4052-883f-1987b7ca0067)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.074966 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac91dd23-e4e5-4073-af76-d760dfdd1adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fz5f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.097490 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.097539 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.097549 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.097565 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.097573 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:20Z","lastTransitionTime":"2026-02-14T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.200350 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.200378 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.200388 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.200401 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.200410 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:20Z","lastTransitionTime":"2026-02-14T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.302673 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.302705 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.302714 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.302726 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.302736 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:20Z","lastTransitionTime":"2026-02-14T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.404789 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.404852 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.404869 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.404888 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.404901 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:20Z","lastTransitionTime":"2026-02-14T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.507022 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.507056 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.507065 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.507079 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.507088 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:20Z","lastTransitionTime":"2026-02-14T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.521233 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jz5f5_0b01cc02-2e62-46e1-b07d-b3c0bff2b050/kube-multus/0.log" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.521283 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jz5f5" event={"ID":"0b01cc02-2e62-46e1-b07d-b3c0bff2b050","Type":"ContainerStarted","Data":"58ce2d3b3613f2f1c1ec36d61edb9bb614e5c1cdf37132359c08ef1f142d8bb0"} Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.533471 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.544960 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.555079 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfcec17d-4d9a-4cdc-a86c-d05e21bedefb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee16aacae98d845ea68a860c26436a43d376f74ea7abd4396984fa63a836cb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc073ab93f3609414e270a023f8d2bad18f489eb6b37ac9a15cd6bf8b4dfbbf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2411332719f59b693409f1e2c9a10c280a102c11216e67d8eea8bfa8bdfe809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.565710 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.575960 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.588272 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.598210 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.606910 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f26819f5-4042-4943-b973-dba56706542e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7f9f319d3a100ef0c94c354ea3722a74d3541a399fe8ccedc4185d34934cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae046a4c4d0f2475bf13ed37f1f9c2029c5b73d436b31705e18ed030f7fc4b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f5pqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.609488 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.609521 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.609530 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.609543 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.609553 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:20Z","lastTransitionTime":"2026-02-14T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.617204 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.628252 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.638286 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.646638 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.657203 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58ce2d3b3613f2f1c1ec36d61edb9bb614e5c1cdf37132359c08ef1f142d8bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:11:19Z\\\",\\\"message\\\":\\\"2026-02-14T11:10:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ae7e0868-7f5e-443d-b025-c96c2b00ebce\\\\n2026-02-14T11:10:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ae7e0868-7f5e-443d-b025-c96c2b00ebce to /host/opt/cni/bin/\\\\n2026-02-14T11:10:33Z [verbose] multus-daemon started\\\\n2026-02-14T11:10:33Z [verbose] Readiness Indicator file check\\\\n2026-02-14T11:11:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.666185 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.675937 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.695724 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"message\\\":\\\"ift-kube-scheduler-crc openshift-multus/network-metrics-daemon-fz5f2 openshift-network-operator/iptables-alerter-4ln5h openshift-network-diagnostics/network-check-target-xd92c openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI0214 11:10:57.697651 6461 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI0214 11:10:57.697673 6461 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697684 6461 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697702 6461 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0214 11:10:57.697710 6461 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0214 11:10:57.697715 6461 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697733 6461 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0214 11:10:57.697811 6461 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2gwlw_openshift-ovn-kubernetes(1ca5acc0-8b15-4052-883f-1987b7ca0067)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.707075 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac91dd23-e4e5-4073-af76-d760dfdd1adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fz5f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:20Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.711518 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.711551 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.711564 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.711581 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.711593 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:20Z","lastTransitionTime":"2026-02-14T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.813234 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.813262 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.813271 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.813283 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.813292 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:20Z","lastTransitionTime":"2026-02-14T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.835784 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:20 crc kubenswrapper[4904]: E0214 11:11:20.835924 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.836771 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 11:34:10.108493766 +0000 UTC Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.915717 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.915775 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.915788 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.915807 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:20 crc kubenswrapper[4904]: I0214 11:11:20.915819 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:20Z","lastTransitionTime":"2026-02-14T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.022287 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.022341 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.022361 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.022383 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.022399 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:21Z","lastTransitionTime":"2026-02-14T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.124507 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.124545 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.124556 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.124573 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.124584 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:21Z","lastTransitionTime":"2026-02-14T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.226889 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.226931 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.226943 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.226960 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.226971 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:21Z","lastTransitionTime":"2026-02-14T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.333948 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.333982 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.333992 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.334007 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.334036 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:21Z","lastTransitionTime":"2026-02-14T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.436306 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.436355 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.436364 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.436378 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.436387 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:21Z","lastTransitionTime":"2026-02-14T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.537929 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.537954 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.537962 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.537975 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.537983 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:21Z","lastTransitionTime":"2026-02-14T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.639830 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.639874 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.639885 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.639901 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.639913 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:21Z","lastTransitionTime":"2026-02-14T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.741353 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.741618 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.741704 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.741787 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.741867 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:21Z","lastTransitionTime":"2026-02-14T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.836248 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.836310 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.836356 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:21 crc kubenswrapper[4904]: E0214 11:11:21.836496 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:21 crc kubenswrapper[4904]: E0214 11:11:21.836545 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:21 crc kubenswrapper[4904]: E0214 11:11:21.836645 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.836884 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 17:42:05.7179953 +0000 UTC Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.843903 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.843922 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.843932 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.843944 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.843954 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:21Z","lastTransitionTime":"2026-02-14T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.946827 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.947092 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.947188 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.947270 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:21 crc kubenswrapper[4904]: I0214 11:11:21.947355 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:21Z","lastTransitionTime":"2026-02-14T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.050454 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.050491 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.050501 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.050514 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.050525 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:22Z","lastTransitionTime":"2026-02-14T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.152379 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.152411 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.152423 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.152439 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.152450 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:22Z","lastTransitionTime":"2026-02-14T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.254802 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.254846 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.254856 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.254868 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.254878 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:22Z","lastTransitionTime":"2026-02-14T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.357046 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.357080 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.357090 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.357106 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.357116 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:22Z","lastTransitionTime":"2026-02-14T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.458976 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.459000 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.459008 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.459021 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.459029 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:22Z","lastTransitionTime":"2026-02-14T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.561156 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.561182 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.561190 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.561202 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.561210 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:22Z","lastTransitionTime":"2026-02-14T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.663399 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.663447 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.663463 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.663479 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.663492 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:22Z","lastTransitionTime":"2026-02-14T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.765730 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.765762 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.765770 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.765784 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.765794 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:22Z","lastTransitionTime":"2026-02-14T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.835791 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:22 crc kubenswrapper[4904]: E0214 11:11:22.835936 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.837947 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 04:45:38.152094429 +0000 UTC Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.867918 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.867940 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.867949 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.867961 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.867969 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:22Z","lastTransitionTime":"2026-02-14T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.970752 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.970774 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.970783 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.970797 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:22 crc kubenswrapper[4904]: I0214 11:11:22.970807 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:22Z","lastTransitionTime":"2026-02-14T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.073404 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.073434 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.073446 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.073462 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.073475 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:23Z","lastTransitionTime":"2026-02-14T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.175690 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.175744 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.175762 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.175782 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.175796 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:23Z","lastTransitionTime":"2026-02-14T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.278141 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.278180 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.278191 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.278204 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.278213 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:23Z","lastTransitionTime":"2026-02-14T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.380441 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.380485 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.380498 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.380515 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.380528 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:23Z","lastTransitionTime":"2026-02-14T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.482488 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.482568 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.482590 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.482651 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.482676 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:23Z","lastTransitionTime":"2026-02-14T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.585388 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.585421 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.585429 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.585443 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.585453 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:23Z","lastTransitionTime":"2026-02-14T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.688067 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.688123 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.688144 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.688167 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.688184 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:23Z","lastTransitionTime":"2026-02-14T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.790288 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.790592 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.790679 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.790784 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.790909 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:23Z","lastTransitionTime":"2026-02-14T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.835894 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.835944 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:23 crc kubenswrapper[4904]: E0214 11:11:23.835989 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.835898 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:23 crc kubenswrapper[4904]: E0214 11:11:23.836084 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:23 crc kubenswrapper[4904]: E0214 11:11:23.836133 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.838984 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 04:55:06.429417598 +0000 UTC Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.893004 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.893028 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.893036 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.893048 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.893056 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:23Z","lastTransitionTime":"2026-02-14T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.995779 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.995852 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.995867 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.995884 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:23 crc kubenswrapper[4904]: I0214 11:11:23.995897 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:23Z","lastTransitionTime":"2026-02-14T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.099078 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.099123 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.099133 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.099169 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.099183 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:24Z","lastTransitionTime":"2026-02-14T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.202041 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.202080 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.202093 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.202112 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.202124 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:24Z","lastTransitionTime":"2026-02-14T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.304422 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.304463 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.304473 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.304487 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.304499 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:24Z","lastTransitionTime":"2026-02-14T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.406990 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.407023 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.407031 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.407044 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.407053 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:24Z","lastTransitionTime":"2026-02-14T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.509723 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.509776 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.509792 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.509810 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.509824 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:24Z","lastTransitionTime":"2026-02-14T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.612895 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.612970 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.612988 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.613012 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.613031 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:24Z","lastTransitionTime":"2026-02-14T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.716106 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.716189 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.716212 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.716241 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.716262 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:24Z","lastTransitionTime":"2026-02-14T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.820235 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.820295 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.820309 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.820329 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.820348 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:24Z","lastTransitionTime":"2026-02-14T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.836071 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:24 crc kubenswrapper[4904]: E0214 11:11:24.836221 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.840110 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 06:55:36.02773795 +0000 UTC Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.923197 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.923249 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.923264 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.923283 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:24 crc kubenswrapper[4904]: I0214 11:11:24.923297 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:24Z","lastTransitionTime":"2026-02-14T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.026555 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.026627 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.026651 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.026681 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.026699 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:25Z","lastTransitionTime":"2026-02-14T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.130607 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.130694 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.130717 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.130757 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.130787 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:25Z","lastTransitionTime":"2026-02-14T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.233133 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.233369 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.233429 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.233492 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.233546 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:25Z","lastTransitionTime":"2026-02-14T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.337503 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.337931 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.338025 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.338129 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.338227 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:25Z","lastTransitionTime":"2026-02-14T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.442069 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.442123 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.442136 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.442158 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.442171 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:25Z","lastTransitionTime":"2026-02-14T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.545469 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.545537 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.545555 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.545585 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.545604 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:25Z","lastTransitionTime":"2026-02-14T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.648628 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.648675 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.648689 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.648707 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.648717 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:25Z","lastTransitionTime":"2026-02-14T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.752338 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.752390 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.752403 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.752424 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.752437 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:25Z","lastTransitionTime":"2026-02-14T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.835945 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.836091 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.836614 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:25 crc kubenswrapper[4904]: E0214 11:11:25.836892 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:25 crc kubenswrapper[4904]: E0214 11:11:25.837078 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.837243 4904 scope.go:117] "RemoveContainer" containerID="f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8" Feb 14 11:11:25 crc kubenswrapper[4904]: E0214 11:11:25.838639 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.840294 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 06:11:29.005224108 +0000 UTC Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.855213 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.855270 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.855290 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.855332 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.855351 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:25Z","lastTransitionTime":"2026-02-14T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.958252 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.958313 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.958329 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.958357 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:25 crc kubenswrapper[4904]: I0214 11:11:25.958383 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:25Z","lastTransitionTime":"2026-02-14T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.060531 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.060564 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.060573 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.060588 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.060598 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:26Z","lastTransitionTime":"2026-02-14T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.163320 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.163382 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.163405 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.163431 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.163449 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:26Z","lastTransitionTime":"2026-02-14T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.266159 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.266224 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.266242 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.266266 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.266283 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:26Z","lastTransitionTime":"2026-02-14T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.368546 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.368591 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.368614 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.368631 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.368642 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:26Z","lastTransitionTime":"2026-02-14T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.470382 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.470422 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.470431 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.470448 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.470457 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:26Z","lastTransitionTime":"2026-02-14T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.530574 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.530621 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.530634 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.530652 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.530666 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:26Z","lastTransitionTime":"2026-02-14T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.542369 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovnkube-controller/2.log" Feb 14 11:11:26 crc kubenswrapper[4904]: E0214 11:11:26.543285 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.545484 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerStarted","Data":"9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70"} Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.545852 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.547652 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.547687 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.547702 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.547721 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.547736 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:26Z","lastTransitionTime":"2026-02-14T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.566496 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: E0214 11:11:26.571356 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.578442 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.578472 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.578481 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.578494 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.578505 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:26Z","lastTransitionTime":"2026-02-14T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.597429 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f26819f5-4042-4943-b973-dba56706542e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7f9f319d3a100ef0c94c354ea3722a74d3541a399fe8ccedc4185d34934cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae046a4c4d0f2475bf13ed37f1f9c2029c5b73d436b31705e18ed030f7fc4b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f5pqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.619605 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: E0214 11:11:26.621642 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.628087 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.628112 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.628121 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.628136 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.628145 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:26Z","lastTransitionTime":"2026-02-14T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.631978 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: E0214 11:11:26.640376 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.642945 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58ce2d3b3613f2f1c1ec36d61edb9bb614e5c1cdf37132359c08ef1f142d8bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:11:19Z\\\",\\\"message\\\":\\\"2026-02-14T11:10:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ae7e0868-7f5e-443d-b025-c96c2b00ebce\\\\n2026-02-14T11:10:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ae7e0868-7f5e-443d-b025-c96c2b00ebce to /host/opt/cni/bin/\\\\n2026-02-14T11:10:33Z [verbose] multus-daemon started\\\\n2026-02-14T11:10:33Z [verbose] Readiness Indicator file check\\\\n2026-02-14T11:11:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.644197 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.644219 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.644241 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.644256 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.644266 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:26Z","lastTransitionTime":"2026-02-14T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.652451 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: E0214 11:11:26.655967 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: E0214 11:11:26.656073 4904 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.657766 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.657800 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.657809 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.657823 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.657850 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:26Z","lastTransitionTime":"2026-02-14T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.666054 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.676533 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.686804 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.709276 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"message\\\":\\\"ift-kube-scheduler-crc openshift-multus/network-metrics-daemon-fz5f2 openshift-network-operator/iptables-alerter-4ln5h openshift-network-diagnostics/network-check-target-xd92c openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI0214 11:10:57.697651 6461 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI0214 11:10:57.697673 6461 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697684 6461 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697702 6461 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0214 11:10:57.697710 6461 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0214 11:10:57.697715 6461 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697733 6461 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0214 11:10:57.697811 6461 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:11:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.719325 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac91dd23-e4e5-4073-af76-d760dfdd1adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fz5f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.732467 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfcec17d-4d9a-4cdc-a86c-d05e21bedefb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee16aacae98d845ea68a860c26436a43d376f74ea7abd4396984fa63a836cb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc073ab93f3609414e270a023f8d2bad18f489eb6b37ac9a15cd6bf8b4dfbbf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2411332719f59b693409f1e2c9a10c280a102c11216e67d8eea8bfa8bdfe809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.744304 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.756517 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.762178 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.762202 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.762211 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.762226 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.762237 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:26Z","lastTransitionTime":"2026-02-14T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.775391 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.787441 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.798156 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:26Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.835423 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:26 crc kubenswrapper[4904]: E0214 11:11:26.835562 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.840766 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 21:12:57.939198844 +0000 UTC Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.864241 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.864272 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.864281 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.864293 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.864301 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:26Z","lastTransitionTime":"2026-02-14T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.966039 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.966074 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.966083 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.966095 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:26 crc kubenswrapper[4904]: I0214 11:11:26.966104 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:26Z","lastTransitionTime":"2026-02-14T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.068090 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.068130 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.068142 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.068157 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.068167 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:27Z","lastTransitionTime":"2026-02-14T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.170443 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.170481 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.170491 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.170513 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.170524 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:27Z","lastTransitionTime":"2026-02-14T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.272990 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.273058 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.273081 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.273110 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.273128 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:27Z","lastTransitionTime":"2026-02-14T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.376512 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.376572 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.376590 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.376618 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.376636 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:27Z","lastTransitionTime":"2026-02-14T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.478874 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.478921 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.478938 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.478960 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.478977 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:27Z","lastTransitionTime":"2026-02-14T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.550234 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovnkube-controller/3.log" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.551400 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovnkube-controller/2.log" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.554373 4904 generic.go:334] "Generic (PLEG): container finished" podID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerID="9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70" exitCode=1 Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.554426 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerDied","Data":"9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70"} Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.554474 4904 scope.go:117] "RemoveContainer" containerID="f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.555695 4904 scope.go:117] "RemoveContainer" containerID="9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70" Feb 14 11:11:27 crc kubenswrapper[4904]: E0214 11:11:27.556109 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2gwlw_openshift-ovn-kubernetes(1ca5acc0-8b15-4052-883f-1987b7ca0067)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.581705 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.581774 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.581792 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.581818 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.581871 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:27Z","lastTransitionTime":"2026-02-14T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.587360 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:27Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.605460 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f26819f5-4042-4943-b973-dba56706542e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7f9f319d3a100ef0c94c354ea3722a74d3541a399fe8ccedc4185d34934cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae046a4c4d0f2475bf13ed37f1f9c2029c5b73d436b31705e18ed030f7fc4b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f5pqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:27Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.619258 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:27Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.636152 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:27Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.654749 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58ce2d3b3613f2f1c1ec36d61edb9bb614e5c1cdf37132359c08ef1f142d8bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:11:19Z\\\",\\\"message\\\":\\\"2026-02-14T11:10:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ae7e0868-7f5e-443d-b025-c96c2b00ebce\\\\n2026-02-14T11:10:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ae7e0868-7f5e-443d-b025-c96c2b00ebce to /host/opt/cni/bin/\\\\n2026-02-14T11:10:33Z [verbose] multus-daemon started\\\\n2026-02-14T11:10:33Z [verbose] Readiness Indicator file check\\\\n2026-02-14T11:11:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:27Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.671697 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:27Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.685359 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.685413 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.685431 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.685455 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.685475 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:27Z","lastTransitionTime":"2026-02-14T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.688085 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:27Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.707476 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:27Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.721903 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:27Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.753406 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f68977c862a6cb28137297f4ca3438af4981cbb20b7941f1a92667c953cc9af8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"message\\\":\\\"ift-kube-scheduler-crc openshift-multus/network-metrics-daemon-fz5f2 openshift-network-operator/iptables-alerter-4ln5h openshift-network-diagnostics/network-check-target-xd92c openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI0214 11:10:57.697651 6461 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI0214 11:10:57.697673 6461 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697684 6461 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697702 6461 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI0214 11:10:57.697710 6461 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI0214 11:10:57.697715 6461 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0214 11:10:57.697733 6461 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0214 11:10:57.697811 6461 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"mns:[] Mutations:[{Column:policies Mutator:insert Value:{GoSet:[{GoUUID:a5a72d02-1a0f-4f7f-a8c5-6923a1c4274a}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:11:26.718473 6856 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI0214 11:11:26.718519 6856 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI0214 11:11:26.718553 6856 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI0214 11:11:26.718602 6856 factory.go:1336] Added *v1.Node event handler 7\\\\nI0214 11:11:26.718638 6856 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0214 11:11:26.718901 6856 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0214 11:11:26.718984 6856 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0214 11:11:26.719028 6856 ovnkube.go:599] Stopped ovnkube\\\\nI0214 11:11:26.719063 6856 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0214 11:11:26.719130 6856 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:11:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:27Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.768268 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac91dd23-e4e5-4073-af76-d760dfdd1adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fz5f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:27Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.784322 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfcec17d-4d9a-4cdc-a86c-d05e21bedefb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee16aacae98d845ea68a860c26436a43d376f74ea7abd4396984fa63a836cb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc073ab93f3609414e270a023f8d2bad18f489eb6b37ac9a15cd6bf8b4dfbbf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2411332719f59b693409f1e2c9a10c280a102c11216e67d8eea8bfa8bdfe809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:27Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.788754 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.788817 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.788883 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.788916 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.788939 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:27Z","lastTransitionTime":"2026-02-14T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.802112 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:27Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.819173 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:27Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.836401 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.836438 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.836450 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:27 crc kubenswrapper[4904]: E0214 11:11:27.837002 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:27 crc kubenswrapper[4904]: E0214 11:11:27.837067 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:27 crc kubenswrapper[4904]: E0214 11:11:27.837218 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.840907 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 02:16:25.462106392 +0000 UTC Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.843159 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:27Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.864036 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:27Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.883632 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:27Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.891896 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.891954 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.891979 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.892007 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.892027 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:27Z","lastTransitionTime":"2026-02-14T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.994674 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.994898 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.994992 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.995058 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:27 crc kubenswrapper[4904]: I0214 11:11:27.995118 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:27Z","lastTransitionTime":"2026-02-14T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.097333 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.097408 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.097429 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.097454 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.097472 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:28Z","lastTransitionTime":"2026-02-14T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.200691 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.200757 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.200774 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.200796 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.200813 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:28Z","lastTransitionTime":"2026-02-14T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.303291 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.303346 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.303361 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.303384 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.303402 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:28Z","lastTransitionTime":"2026-02-14T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.406201 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.406240 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.406250 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.406265 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.406277 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:28Z","lastTransitionTime":"2026-02-14T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.508123 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.508165 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.508176 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.508191 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.508202 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:28Z","lastTransitionTime":"2026-02-14T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.558491 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovnkube-controller/3.log" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.562067 4904 scope.go:117] "RemoveContainer" containerID="9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70" Feb 14 11:11:28 crc kubenswrapper[4904]: E0214 11:11:28.562209 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2gwlw_openshift-ovn-kubernetes(1ca5acc0-8b15-4052-883f-1987b7ca0067)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.576767 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:28Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.589118 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f26819f5-4042-4943-b973-dba56706542e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7f9f319d3a100ef0c94c354ea3722a74d3541a399fe8ccedc4185d34934cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae046a4c4d0f2475bf13ed37f1f9c2029c5b73d436b31705e18ed030f7fc4b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f5pqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:28Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.600265 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:28Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.610588 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.610657 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.610668 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.610686 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.610697 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:28Z","lastTransitionTime":"2026-02-14T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.612351 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58ce2d3b3613f2f1c1ec36d61edb9bb614e5c1cdf37132359c08ef1f142d8bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:11:19Z\\\",\\\"message\\\":\\\"2026-02-14T11:10:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ae7e0868-7f5e-443d-b025-c96c2b00ebce\\\\n2026-02-14T11:10:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ae7e0868-7f5e-443d-b025-c96c2b00ebce to /host/opt/cni/bin/\\\\n2026-02-14T11:10:33Z [verbose] multus-daemon started\\\\n2026-02-14T11:10:33Z [verbose] Readiness Indicator file check\\\\n2026-02-14T11:11:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:28Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.621334 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:28Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.633282 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:28Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.644371 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:28Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.653226 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:28Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.668625 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"mns:[] Mutations:[{Column:policies Mutator:insert Value:{GoSet:[{GoUUID:a5a72d02-1a0f-4f7f-a8c5-6923a1c4274a}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:11:26.718473 6856 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI0214 11:11:26.718519 6856 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI0214 11:11:26.718553 6856 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI0214 11:11:26.718602 6856 factory.go:1336] Added *v1.Node event handler 7\\\\nI0214 11:11:26.718638 6856 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0214 11:11:26.718901 6856 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0214 11:11:26.718984 6856 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0214 11:11:26.719028 6856 ovnkube.go:599] Stopped ovnkube\\\\nI0214 11:11:26.719063 6856 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0214 11:11:26.719130 6856 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:11:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2gwlw_openshift-ovn-kubernetes(1ca5acc0-8b15-4052-883f-1987b7ca0067)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:28Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.679351 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac91dd23-e4e5-4073-af76-d760dfdd1adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fz5f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:28Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.687295 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:28Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.697117 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:28Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.708325 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:28Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.712376 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.712402 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.712413 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.712429 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.712440 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:28Z","lastTransitionTime":"2026-02-14T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.720674 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:28Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.734669 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:28Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.749074 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:28Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.759951 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfcec17d-4d9a-4cdc-a86c-d05e21bedefb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee16aacae98d845ea68a860c26436a43d376f74ea7abd4396984fa63a836cb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc073ab93f3609414e270a023f8d2bad18f489eb6b37ac9a15cd6bf8b4dfbbf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2411332719f59b693409f1e2c9a10c280a102c11216e67d8eea8bfa8bdfe809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:28Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.815806 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.815857 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.815866 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.815881 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.815893 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:28Z","lastTransitionTime":"2026-02-14T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.836212 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:28 crc kubenswrapper[4904]: E0214 11:11:28.836336 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.841538 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 07:34:16.305729197 +0000 UTC Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.918614 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.918661 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.918674 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.918692 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:28 crc kubenswrapper[4904]: I0214 11:11:28.918707 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:28Z","lastTransitionTime":"2026-02-14T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.021771 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.021863 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.021878 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.021911 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.021927 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:29Z","lastTransitionTime":"2026-02-14T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.124960 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.125057 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.125082 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.125123 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.125148 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:29Z","lastTransitionTime":"2026-02-14T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.228788 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.228858 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.228869 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.228887 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.228902 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:29Z","lastTransitionTime":"2026-02-14T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.333028 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.333085 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.333099 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.333117 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.333130 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:29Z","lastTransitionTime":"2026-02-14T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.436316 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.436361 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.436393 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.436413 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.436423 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:29Z","lastTransitionTime":"2026-02-14T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.539915 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.539982 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.539995 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.540014 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.540027 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:29Z","lastTransitionTime":"2026-02-14T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.642959 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.643007 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.643019 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.643035 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.643052 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:29Z","lastTransitionTime":"2026-02-14T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.745277 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.745311 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.745322 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.745338 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.745350 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:29Z","lastTransitionTime":"2026-02-14T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.835452 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.835452 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.835577 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:29 crc kubenswrapper[4904]: E0214 11:11:29.835777 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:29 crc kubenswrapper[4904]: E0214 11:11:29.835897 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:29 crc kubenswrapper[4904]: E0214 11:11:29.836000 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.843026 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 19:30:42.117858422 +0000 UTC Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.847267 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.847328 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.847347 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.847369 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.847415 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:29Z","lastTransitionTime":"2026-02-14T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.848061 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:29Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.858757 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f26819f5-4042-4943-b973-dba56706542e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7f9f319d3a100ef0c94c354ea3722a74d3541a399fe8ccedc4185d34934cbfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae046a4c4d0f2475bf13ed37f1f9c2029c5b73d436b31705e18ed030f7fc4b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vt5kv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f5pqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:29Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.871367 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7790d276-2506-4641-8c22-67e6441417f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d81e847c76669f3de1cb2e56aa6839d578864fe80267a610cd490ca2cdd740a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqg5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-tnf2t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:29Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.883897 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jz5f5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b01cc02-2e62-46e1-b07d-b3c0bff2b050\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58ce2d3b3613f2f1c1ec36d61edb9bb614e5c1cdf37132359c08ef1f142d8bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:11:19Z\\\",\\\"message\\\":\\\"2026-02-14T11:10:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ae7e0868-7f5e-443d-b025-c96c2b00ebce\\\\n2026-02-14T11:10:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ae7e0868-7f5e-443d-b025-c96c2b00ebce to /host/opt/cni/bin/\\\\n2026-02-14T11:10:33Z [verbose] multus-daemon started\\\\n2026-02-14T11:10:33Z [verbose] Readiness Indicator file check\\\\n2026-02-14T11:11:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ccqlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jz5f5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:29Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.897237 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrwvd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04ac70df-4f95-4a9e-82d8-164ae3a71623\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c130f60f2db0c1c94e677b0b5df7b21dc455f1bba3e72c593d135f644b1a856e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxbls\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrwvd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:29Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.909899 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb23a7f689fb23f02f028ef1dbf4f71086f3c4f0377d50050a27c53d53f773f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:29Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.928981 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:29Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.941115 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77a1621edaadb02939bc14e064170e518a41e948811c1b550b27eb2605e4beda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:29Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.949911 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.950017 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.950029 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.950042 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.950052 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:29Z","lastTransitionTime":"2026-02-14T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.964021 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ca5acc0-8b15-4052-883f-1987b7ca0067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-14T11:11:26Z\\\",\\\"message\\\":\\\"mns:[] Mutations:[{Column:policies Mutator:insert Value:{GoSet:[{GoUUID:a5a72d02-1a0f-4f7f-a8c5-6923a1c4274a}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0214 11:11:26.718473 6856 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI0214 11:11:26.718519 6856 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI0214 11:11:26.718553 6856 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI0214 11:11:26.718602 6856 factory.go:1336] Added *v1.Node event handler 7\\\\nI0214 11:11:26.718638 6856 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0214 11:11:26.718901 6856 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0214 11:11:26.718984 6856 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0214 11:11:26.719028 6856 ovnkube.go:599] Stopped ovnkube\\\\nI0214 11:11:26.719063 6856 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0214 11:11:26.719130 6856 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:11:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2gwlw_openshift-ovn-kubernetes(1ca5acc0-8b15-4052-883f-1987b7ca0067)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rghsz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2gwlw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:29Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.973760 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac91dd23-e4e5-4073-af76-d760dfdd1adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkv4r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fz5f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:29Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.982284 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zjhq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95999cbb-5c5a-4dd9-afbd-6472f52bf66b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89a396273f7f461243397d933a184701824bd2a04c3e9277968561d4094a2fe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vv4sg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zjhq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:29Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:29 crc kubenswrapper[4904]: I0214 11:11:29.993139 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a836ad5f8efa90979f03f21f0400b0270abecd7471c839fb9112890b850816b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed4db3596ccceb2a6b770ef88b20e26d334acd9675c3b0c7f8af7830c7d844d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:29Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.003442 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:30Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.015120 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-65flh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4edce510-9ef8-498a-8b5c-9b9eb41847eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://709723bc3ff1b67ea6e3b31d7fe1d759b3408977f6ae5c5cfc0b011e84ed2593\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fe42648d156c3d016c229b300623c277085d3a8a767c28fa9e9f15b0ac7d50e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec2ec5d28861084cfc82b09e1f20dcd0eff1e6cbc00930418101c7870b30c12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04269d857356ee101ba9b590d83e3834c302fac257485d3b66342d0ab05c9668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1294b3bd1b15e4932fa52a385471f3a899fecfc144aa84b2dc326ef1ba0ae9dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://406de114ef87c530869c947cce3487494a49762ab06b80c6077732cb7850d55a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b7ca35b88174331d04098a6d0978338b5b6e758e72ec3f47aeedb3d1dd0a86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4rxmn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-65flh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:30Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.029236 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed38d346-9528-40d4-96fd-ad2476654244\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-14T11:10:28Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0214 11:10:23.240093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0214 11:10:23.240674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2777043737/tls.crt::/tmp/serving-cert-2777043737/tls.key\\\\\\\"\\\\nI0214 11:10:28.931121 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0214 11:10:28.934133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0214 11:10:28.934156 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0214 11:10:28.934181 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0214 11:10:28.934188 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0214 11:10:28.946176 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0214 11:10:28.946215 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946221 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0214 11:10:28.946227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0214 11:10:28.946274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0214 11:10:28.946283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0214 11:10:28.946287 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0214 11:10:28.946757 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0214 11:10:28.949602 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:30Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.046094 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dea1410-fb93-4fea-b169-e9bbcc176572\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee40d4934189e317a3b6cb65d0f555d8d140d727d3f339b1cdff509d5c881db7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fac5236d22f805edc365ac96f3e508780780f3b0b2d9be80a63ccb62c09def7b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a314cb926d6a42801e8b556bcb55e012973a00ca160d60dd86171ab8e123768f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:30Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.053340 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.053373 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.053385 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.053397 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.053406 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:30Z","lastTransitionTime":"2026-02-14T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.059698 4904 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfcec17d-4d9a-4cdc-a86c-d05e21bedefb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-14T11:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee16aacae98d845ea68a860c26436a43d376f74ea7abd4396984fa63a836cb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc073ab93f3609414e270a023f8d2bad18f489eb6b37ac9a15cd6bf8b4dfbbf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2411332719f59b693409f1e2c9a10c280a102c11216e67d8eea8bfa8bdfe809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-14T11:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a520f7827a51949c05e14e9116df3c5e577cc5907516bf6785a3d3270bdc04e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-14T11:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-14T11:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-14T11:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:30Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.155610 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.155658 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.155669 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.155726 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.155745 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:30Z","lastTransitionTime":"2026-02-14T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.258092 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.258132 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.258142 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.258191 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.258273 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:30Z","lastTransitionTime":"2026-02-14T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.360083 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.360124 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.360136 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.360152 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.360164 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:30Z","lastTransitionTime":"2026-02-14T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.462555 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.462584 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.462593 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.462605 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.462614 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:30Z","lastTransitionTime":"2026-02-14T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.564877 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.564905 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.564913 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.564925 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.564934 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:30Z","lastTransitionTime":"2026-02-14T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.666691 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.666751 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.666763 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.666780 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.666812 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:30Z","lastTransitionTime":"2026-02-14T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.769232 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.769267 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.769278 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.769293 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.769303 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:30Z","lastTransitionTime":"2026-02-14T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.836020 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:30 crc kubenswrapper[4904]: E0214 11:11:30.836175 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.843454 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 09:03:17.222747618 +0000 UTC Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.871906 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.871943 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.871951 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.871963 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.871972 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:30Z","lastTransitionTime":"2026-02-14T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.974248 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.974278 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.974287 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.974299 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:30 crc kubenswrapper[4904]: I0214 11:11:30.974308 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:30Z","lastTransitionTime":"2026-02-14T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.076326 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.076367 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.076378 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.076392 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.076401 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:31Z","lastTransitionTime":"2026-02-14T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.179052 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.179095 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.179106 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.179122 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.179132 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:31Z","lastTransitionTime":"2026-02-14T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.281651 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.281694 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.281705 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.281722 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.281738 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:31Z","lastTransitionTime":"2026-02-14T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.383595 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.383633 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.383641 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.383656 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.383667 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:31Z","lastTransitionTime":"2026-02-14T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.486290 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.486350 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.486369 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.486394 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.486411 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:31Z","lastTransitionTime":"2026-02-14T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.588221 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.588248 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.588255 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.588267 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.588276 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:31Z","lastTransitionTime":"2026-02-14T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.690957 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.690987 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.690997 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.691010 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.691019 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:31Z","lastTransitionTime":"2026-02-14T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.792962 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.793002 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.793015 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.793031 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.793042 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:31Z","lastTransitionTime":"2026-02-14T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.836014 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.836014 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:31 crc kubenswrapper[4904]: E0214 11:11:31.836119 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.836214 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:31 crc kubenswrapper[4904]: E0214 11:11:31.836307 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:31 crc kubenswrapper[4904]: E0214 11:11:31.836527 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.844043 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 01:24:10.278919226 +0000 UTC Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.895469 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.895523 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.895540 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.895562 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.895580 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:31Z","lastTransitionTime":"2026-02-14T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.998752 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.998799 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.998817 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.998867 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:31 crc kubenswrapper[4904]: I0214 11:11:31.998888 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:31Z","lastTransitionTime":"2026-02-14T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.101822 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.101920 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.101941 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.101968 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.101988 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:32Z","lastTransitionTime":"2026-02-14T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.205096 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.205170 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.205186 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.205208 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.205223 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:32Z","lastTransitionTime":"2026-02-14T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.307310 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.307365 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.307382 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.307402 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.307417 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:32Z","lastTransitionTime":"2026-02-14T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.409644 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.409678 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.409689 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.409708 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.409723 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:32Z","lastTransitionTime":"2026-02-14T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.512199 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.512273 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.512295 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.512323 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.512342 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:32Z","lastTransitionTime":"2026-02-14T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.614908 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.614962 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.614980 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.615006 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.615023 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:32Z","lastTransitionTime":"2026-02-14T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.717690 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.717749 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.717766 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.717789 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.717806 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:32Z","lastTransitionTime":"2026-02-14T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.820504 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.820569 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.820583 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.820603 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.820616 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:32Z","lastTransitionTime":"2026-02-14T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.836326 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:32 crc kubenswrapper[4904]: E0214 11:11:32.836500 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.844589 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 14:37:23.565127298 +0000 UTC Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.922855 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.922886 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.922918 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.922933 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:32 crc kubenswrapper[4904]: I0214 11:11:32.922961 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:32Z","lastTransitionTime":"2026-02-14T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.025572 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.025607 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.025647 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.025665 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.025675 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:33Z","lastTransitionTime":"2026-02-14T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.134153 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.134249 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.134298 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.134336 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.134362 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:33Z","lastTransitionTime":"2026-02-14T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.237153 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.237199 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.237208 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.237223 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.237232 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:33Z","lastTransitionTime":"2026-02-14T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.340150 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.340203 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.340219 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.340239 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.340252 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:33Z","lastTransitionTime":"2026-02-14T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.442541 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.442581 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.442593 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.442608 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.442618 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:33Z","lastTransitionTime":"2026-02-14T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.544935 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.544971 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.544983 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.545000 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.545009 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:33Z","lastTransitionTime":"2026-02-14T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.646812 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.646878 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.646889 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.646903 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.646916 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:33Z","lastTransitionTime":"2026-02-14T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.749642 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.749683 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.749692 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.749707 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.749717 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:33Z","lastTransitionTime":"2026-02-14T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.755959 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:11:33 crc kubenswrapper[4904]: E0214 11:11:33.756108 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.75609254 +0000 UTC m=+148.568857191 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.835857 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.835912 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.835955 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:33 crc kubenswrapper[4904]: E0214 11:11:33.836038 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:33 crc kubenswrapper[4904]: E0214 11:11:33.836181 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:33 crc kubenswrapper[4904]: E0214 11:11:33.836410 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.845492 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 06:46:30.613048042 +0000 UTC Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.852373 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.852409 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.852419 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.852435 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.852444 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:33Z","lastTransitionTime":"2026-02-14T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.856798 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.856865 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.856887 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.856920 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:33 crc kubenswrapper[4904]: E0214 11:11:33.857014 4904 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 14 11:11:33 crc kubenswrapper[4904]: E0214 11:11:33.857056 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.857041456 +0000 UTC m=+148.669806117 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 14 11:11:33 crc kubenswrapper[4904]: E0214 11:11:33.857108 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 14 11:11:33 crc kubenswrapper[4904]: E0214 11:11:33.857113 4904 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 14 11:11:33 crc kubenswrapper[4904]: E0214 11:11:33.857135 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 14 11:11:33 crc kubenswrapper[4904]: E0214 11:11:33.857179 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 14 11:11:33 crc kubenswrapper[4904]: E0214 11:11:33.857127 4904 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 14 11:11:33 crc kubenswrapper[4904]: E0214 11:11:33.857204 4904 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:11:33 crc kubenswrapper[4904]: E0214 11:11:33.857237 4904 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:11:33 crc kubenswrapper[4904]: E0214 11:11:33.857220 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.85718569 +0000 UTC m=+148.669950401 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 14 11:11:33 crc kubenswrapper[4904]: E0214 11:11:33.857311 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.857284312 +0000 UTC m=+148.670049073 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:11:33 crc kubenswrapper[4904]: E0214 11:11:33.857392 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.857370826 +0000 UTC m=+148.670135607 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.954560 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.954588 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.954597 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.954610 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:33 crc kubenswrapper[4904]: I0214 11:11:33.954618 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:33Z","lastTransitionTime":"2026-02-14T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.058147 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.058183 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.058195 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.058211 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.058221 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:34Z","lastTransitionTime":"2026-02-14T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.160735 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.160774 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.160783 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.160797 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.160805 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:34Z","lastTransitionTime":"2026-02-14T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.262798 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.262855 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.262865 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.262879 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.262887 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:34Z","lastTransitionTime":"2026-02-14T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.365081 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.365123 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.365133 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.365148 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.365157 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:34Z","lastTransitionTime":"2026-02-14T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.467584 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.467616 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.467624 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.467637 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.467645 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:34Z","lastTransitionTime":"2026-02-14T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.570188 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.570689 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.570711 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.570738 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.570756 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:34Z","lastTransitionTime":"2026-02-14T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.673288 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.673319 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.673329 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.673364 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.673375 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:34Z","lastTransitionTime":"2026-02-14T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.776039 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.776101 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.776118 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.776143 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.776160 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:34Z","lastTransitionTime":"2026-02-14T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.835940 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:34 crc kubenswrapper[4904]: E0214 11:11:34.836095 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.846242 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 09:12:33.127248034 +0000 UTC Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.879856 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.879903 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.879918 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.879936 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.879950 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:34Z","lastTransitionTime":"2026-02-14T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.982895 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.982955 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.982972 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.983000 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:34 crc kubenswrapper[4904]: I0214 11:11:34.983018 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:34Z","lastTransitionTime":"2026-02-14T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.085248 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.085316 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.085334 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.085367 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.085389 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:35Z","lastTransitionTime":"2026-02-14T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.187889 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.187919 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.187930 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.187945 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.187957 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:35Z","lastTransitionTime":"2026-02-14T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.290353 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.290387 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.290424 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.290442 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.290454 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:35Z","lastTransitionTime":"2026-02-14T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.393435 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.393531 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.393552 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.393577 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.393595 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:35Z","lastTransitionTime":"2026-02-14T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.496193 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.496241 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.496254 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.496266 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.496275 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:35Z","lastTransitionTime":"2026-02-14T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.597923 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.598042 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.598051 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.598063 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.598073 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:35Z","lastTransitionTime":"2026-02-14T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.707107 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.707165 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.707176 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.707193 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.707203 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:35Z","lastTransitionTime":"2026-02-14T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.810186 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.810229 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.810244 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.810264 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.810278 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:35Z","lastTransitionTime":"2026-02-14T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.835816 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.835925 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.836059 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:35 crc kubenswrapper[4904]: E0214 11:11:35.836068 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:35 crc kubenswrapper[4904]: E0214 11:11:35.836256 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:35 crc kubenswrapper[4904]: E0214 11:11:35.836346 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.847084 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 16:35:38.934280411 +0000 UTC Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.854960 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.855077 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.913974 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.914435 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.914795 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.915039 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:35 crc kubenswrapper[4904]: I0214 11:11:35.915207 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:35Z","lastTransitionTime":"2026-02-14T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.018791 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.018885 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.018903 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.018930 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.018947 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:36Z","lastTransitionTime":"2026-02-14T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.122125 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.122172 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.122184 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.122202 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.122212 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:36Z","lastTransitionTime":"2026-02-14T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.224998 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.225066 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.225209 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.225259 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.225284 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:36Z","lastTransitionTime":"2026-02-14T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.328740 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.328809 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.328826 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.328863 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.328878 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:36Z","lastTransitionTime":"2026-02-14T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.431983 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.432058 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.432082 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.432112 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.432132 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:36Z","lastTransitionTime":"2026-02-14T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.535074 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.535129 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.535148 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.535173 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.535191 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:36Z","lastTransitionTime":"2026-02-14T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.637701 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.637791 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.637810 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.637871 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.637897 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:36Z","lastTransitionTime":"2026-02-14T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.710145 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.710397 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.710423 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.710446 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.710463 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:36Z","lastTransitionTime":"2026-02-14T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:36 crc kubenswrapper[4904]: E0214 11:11:36.731251 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:36Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.735177 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.735217 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.735234 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.735255 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.735271 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:36Z","lastTransitionTime":"2026-02-14T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:36 crc kubenswrapper[4904]: E0214 11:11:36.753176 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:36Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.756417 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.756473 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.756485 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.756501 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.756513 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:36Z","lastTransitionTime":"2026-02-14T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:36 crc kubenswrapper[4904]: E0214 11:11:36.771058 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:36Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.774589 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.774655 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.774666 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.774682 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.774716 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:36Z","lastTransitionTime":"2026-02-14T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:36 crc kubenswrapper[4904]: E0214 11:11:36.791316 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:36Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.795820 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.795888 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.795903 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.795936 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.795951 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:36Z","lastTransitionTime":"2026-02-14T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:36 crc kubenswrapper[4904]: E0214 11:11:36.807026 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-14T11:11:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6bfee4df-cb7e-45ab-a7f5-35741705b859\\\",\\\"systemUUID\\\":\\\"8976ab10-78e1-4979-8878-7acc2b1073d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-14T11:11:36Z is after 2025-08-24T17:21:41Z" Feb 14 11:11:36 crc kubenswrapper[4904]: E0214 11:11:36.807265 4904 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.808960 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.808987 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.808997 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.809034 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.809045 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:36Z","lastTransitionTime":"2026-02-14T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.835607 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:36 crc kubenswrapper[4904]: E0214 11:11:36.835719 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.847817 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 16:35:06.719224359 +0000 UTC Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.911514 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.911566 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.911580 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.911593 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:36 crc kubenswrapper[4904]: I0214 11:11:36.911603 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:36Z","lastTransitionTime":"2026-02-14T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.014295 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.014347 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.014363 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.014382 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.014397 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:37Z","lastTransitionTime":"2026-02-14T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.117053 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.117104 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.117120 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.117143 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.117159 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:37Z","lastTransitionTime":"2026-02-14T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.223965 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.224007 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.224019 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.224036 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.224048 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:37Z","lastTransitionTime":"2026-02-14T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.327200 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.327244 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.327258 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.327277 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.327291 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:37Z","lastTransitionTime":"2026-02-14T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.429936 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.429974 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.429985 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.430003 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.430018 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:37Z","lastTransitionTime":"2026-02-14T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.533584 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.533634 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.533653 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.533678 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.533695 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:37Z","lastTransitionTime":"2026-02-14T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.636782 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.636922 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.636945 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.636973 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.636992 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:37Z","lastTransitionTime":"2026-02-14T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.739749 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.739806 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.739817 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.739861 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.739874 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:37Z","lastTransitionTime":"2026-02-14T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.836045 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.836248 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:37 crc kubenswrapper[4904]: E0214 11:11:37.836563 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:37 crc kubenswrapper[4904]: E0214 11:11:37.836734 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.837129 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:37 crc kubenswrapper[4904]: E0214 11:11:37.837309 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.841325 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.841391 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.841417 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.841446 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.841469 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:37Z","lastTransitionTime":"2026-02-14T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.848606 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 03:03:23.075042135 +0000 UTC Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.944581 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.944653 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.944672 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.944703 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:37 crc kubenswrapper[4904]: I0214 11:11:37.944725 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:37Z","lastTransitionTime":"2026-02-14T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.048458 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.048539 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.048560 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.048586 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.048605 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:38Z","lastTransitionTime":"2026-02-14T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.151122 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.151156 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.151165 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.151179 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.151189 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:38Z","lastTransitionTime":"2026-02-14T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.253883 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.253913 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.253921 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.253978 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.253988 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:38Z","lastTransitionTime":"2026-02-14T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.357298 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.357379 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.357398 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.357423 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.357440 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:38Z","lastTransitionTime":"2026-02-14T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.460156 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.460226 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.460250 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.460281 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.460305 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:38Z","lastTransitionTime":"2026-02-14T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.563324 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.563363 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.563374 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.563392 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.563406 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:38Z","lastTransitionTime":"2026-02-14T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.666329 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.666415 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.666437 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.666460 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.666476 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:38Z","lastTransitionTime":"2026-02-14T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.770042 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.770075 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.770087 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.770102 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.770113 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:38Z","lastTransitionTime":"2026-02-14T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.835864 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:38 crc kubenswrapper[4904]: E0214 11:11:38.836217 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.849476 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 06:19:44.133105525 +0000 UTC Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.873263 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.873366 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.873398 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.873428 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.873451 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:38Z","lastTransitionTime":"2026-02-14T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.976106 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.976164 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.976188 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.976216 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:38 crc kubenswrapper[4904]: I0214 11:11:38.976240 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:38Z","lastTransitionTime":"2026-02-14T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.080106 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.080302 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.080344 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.080377 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.080400 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:39Z","lastTransitionTime":"2026-02-14T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.183426 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.183473 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.183484 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.183506 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.183518 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:39Z","lastTransitionTime":"2026-02-14T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.285742 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.285787 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.285799 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.285816 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.285845 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:39Z","lastTransitionTime":"2026-02-14T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.387894 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.387994 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.388013 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.388086 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.388104 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:39Z","lastTransitionTime":"2026-02-14T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.491331 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.491404 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.491428 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.491458 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.491480 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:39Z","lastTransitionTime":"2026-02-14T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.595582 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.595631 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.595654 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.595680 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.595699 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:39Z","lastTransitionTime":"2026-02-14T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.698087 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.698133 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.698141 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.698156 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.698165 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:39Z","lastTransitionTime":"2026-02-14T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.801212 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.801276 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.801288 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.801308 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.801319 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:39Z","lastTransitionTime":"2026-02-14T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.836181 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.836186 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.836281 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:39 crc kubenswrapper[4904]: E0214 11:11:39.836513 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:39 crc kubenswrapper[4904]: E0214 11:11:39.837004 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:39 crc kubenswrapper[4904]: E0214 11:11:39.837181 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.850616 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 18:41:21.116298328 +0000 UTC Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.880616 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-jz5f5" podStartSLOduration=69.880593239 podStartE2EDuration="1m9.880593239s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:11:39.862632105 +0000 UTC m=+90.675396756" watchObservedRunningTime="2026-02-14 11:11:39.880593239 +0000 UTC m=+90.693357900" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.899759 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-zrwvd" podStartSLOduration=69.899735635 podStartE2EDuration="1m9.899735635s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:11:39.88136088 +0000 UTC m=+90.694125541" watchObservedRunningTime="2026-02-14 11:11:39.899735635 +0000 UTC m=+90.712500306" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.906393 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.906740 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.906876 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.906981 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.907072 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:39Z","lastTransitionTime":"2026-02-14T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.955697 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podStartSLOduration=69.955672092 podStartE2EDuration="1m9.955672092s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:11:39.95520449 +0000 UTC m=+90.767969151" watchObservedRunningTime="2026-02-14 11:11:39.955672092 +0000 UTC m=+90.768436763" Feb 14 11:11:39 crc kubenswrapper[4904]: I0214 11:11:39.978540 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=4.978524955 podStartE2EDuration="4.978524955s" podCreationTimestamp="2026-02-14 11:11:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:11:39.97756168 +0000 UTC m=+90.790326341" watchObservedRunningTime="2026-02-14 11:11:39.978524955 +0000 UTC m=+90.791289616" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.002464 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=5.002446037 podStartE2EDuration="5.002446037s" podCreationTimestamp="2026-02-14 11:11:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:11:40.001445901 +0000 UTC m=+90.814210562" watchObservedRunningTime="2026-02-14 11:11:40.002446037 +0000 UTC m=+90.815210708" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.009498 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.009537 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.009546 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.009561 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.009571 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:40Z","lastTransitionTime":"2026-02-14T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.040169 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-zjhq2" podStartSLOduration=70.040152613 podStartE2EDuration="1m10.040152613s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:11:40.017467104 +0000 UTC m=+90.830231765" watchObservedRunningTime="2026-02-14 11:11:40.040152613 +0000 UTC m=+90.852917274" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.091614 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-65flh" podStartSLOduration=70.091596981 podStartE2EDuration="1m10.091596981s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:11:40.071622904 +0000 UTC m=+90.884387575" watchObservedRunningTime="2026-02-14 11:11:40.091596981 +0000 UTC m=+90.904361642" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.106907 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=70.106884595 podStartE2EDuration="1m10.106884595s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:11:40.092529156 +0000 UTC m=+90.905293827" watchObservedRunningTime="2026-02-14 11:11:40.106884595 +0000 UTC m=+90.919649256" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.107610 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=67.107603385 podStartE2EDuration="1m7.107603385s" podCreationTimestamp="2026-02-14 11:10:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:11:40.106757312 +0000 UTC m=+90.919521993" watchObservedRunningTime="2026-02-14 11:11:40.107603385 +0000 UTC m=+90.920368046" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.112301 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.112348 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.112358 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.112374 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.112383 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:40Z","lastTransitionTime":"2026-02-14T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.142137 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=43.142119726 podStartE2EDuration="43.142119726s" podCreationTimestamp="2026-02-14 11:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:11:40.125359883 +0000 UTC m=+90.938124544" watchObservedRunningTime="2026-02-14 11:11:40.142119726 +0000 UTC m=+90.954884387" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.166777 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f5pqm" podStartSLOduration=69.166751396 podStartE2EDuration="1m9.166751396s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:11:40.166279253 +0000 UTC m=+90.979043904" watchObservedRunningTime="2026-02-14 11:11:40.166751396 +0000 UTC m=+90.979516057" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.214571 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.214614 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.214621 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.214638 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.214651 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:40Z","lastTransitionTime":"2026-02-14T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.316752 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.316800 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.316810 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.316827 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.316874 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:40Z","lastTransitionTime":"2026-02-14T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.419468 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.419499 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.419510 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.419527 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.419566 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:40Z","lastTransitionTime":"2026-02-14T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.523220 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.523254 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.523268 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.523287 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.523300 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:40Z","lastTransitionTime":"2026-02-14T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.625921 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.625960 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.625971 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.625986 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.625998 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:40Z","lastTransitionTime":"2026-02-14T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.728889 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.728943 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.728955 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.728974 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.728989 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:40Z","lastTransitionTime":"2026-02-14T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.831687 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.831722 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.831734 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.831753 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.831766 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:40Z","lastTransitionTime":"2026-02-14T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.835660 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:40 crc kubenswrapper[4904]: E0214 11:11:40.835815 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.850819 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 06:44:46.742531772 +0000 UTC Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.933951 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.934006 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.934022 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.934042 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:40 crc kubenswrapper[4904]: I0214 11:11:40.934058 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:40Z","lastTransitionTime":"2026-02-14T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.036444 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.036486 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.036499 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.036516 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.036528 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:41Z","lastTransitionTime":"2026-02-14T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.138460 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.138496 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.138506 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.138520 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.138532 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:41Z","lastTransitionTime":"2026-02-14T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.241100 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.241139 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.241151 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.241171 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.241183 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:41Z","lastTransitionTime":"2026-02-14T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.344152 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.344240 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.344291 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.344314 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.344331 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:41Z","lastTransitionTime":"2026-02-14T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.447279 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.447346 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.447362 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.447381 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.447416 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:41Z","lastTransitionTime":"2026-02-14T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.549657 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.549754 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.549781 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.549808 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.549923 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:41Z","lastTransitionTime":"2026-02-14T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.653394 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.653423 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.653433 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.653445 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.653454 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:41Z","lastTransitionTime":"2026-02-14T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.755447 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.755494 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.755509 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.755526 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.755540 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:41Z","lastTransitionTime":"2026-02-14T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.836487 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.836514 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:41 crc kubenswrapper[4904]: E0214 11:11:41.836711 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.836514 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:41 crc kubenswrapper[4904]: E0214 11:11:41.836828 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:41 crc kubenswrapper[4904]: E0214 11:11:41.837096 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.851282 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 22:59:16.828723753 +0000 UTC Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.858451 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.858502 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.858516 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.858533 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.858547 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:41Z","lastTransitionTime":"2026-02-14T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.961094 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.961117 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.961125 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.961137 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:41 crc kubenswrapper[4904]: I0214 11:11:41.961146 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:41Z","lastTransitionTime":"2026-02-14T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.063251 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.063279 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.063295 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.063309 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.063318 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:42Z","lastTransitionTime":"2026-02-14T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.165879 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.165904 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.165913 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.165925 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.165933 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:42Z","lastTransitionTime":"2026-02-14T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.269085 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.269167 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.269197 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.269225 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.269246 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:42Z","lastTransitionTime":"2026-02-14T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.371771 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.371808 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.371816 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.371830 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.371859 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:42Z","lastTransitionTime":"2026-02-14T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.474600 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.474629 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.474640 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.474656 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.474664 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:42Z","lastTransitionTime":"2026-02-14T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.576978 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.577023 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.577033 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.577049 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.577058 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:42Z","lastTransitionTime":"2026-02-14T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.679757 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.679807 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.679817 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.679846 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.679855 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:42Z","lastTransitionTime":"2026-02-14T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.782595 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.782657 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.782674 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.782697 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.782716 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:42Z","lastTransitionTime":"2026-02-14T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.836001 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:42 crc kubenswrapper[4904]: E0214 11:11:42.836395 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.836589 4904 scope.go:117] "RemoveContainer" containerID="9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70" Feb 14 11:11:42 crc kubenswrapper[4904]: E0214 11:11:42.836745 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2gwlw_openshift-ovn-kubernetes(1ca5acc0-8b15-4052-883f-1987b7ca0067)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.851413 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 20:34:20.951282415 +0000 UTC Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.885488 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.885520 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.885530 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.885545 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.885557 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:42Z","lastTransitionTime":"2026-02-14T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.988165 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.988226 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.988243 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.988272 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:42 crc kubenswrapper[4904]: I0214 11:11:42.988288 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:42Z","lastTransitionTime":"2026-02-14T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.090362 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.090402 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.090414 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.090430 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.090443 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:43Z","lastTransitionTime":"2026-02-14T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.193181 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.193225 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.193235 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.193248 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.193257 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:43Z","lastTransitionTime":"2026-02-14T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.295745 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.295781 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.295789 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.295801 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.295810 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:43Z","lastTransitionTime":"2026-02-14T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.397689 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.397732 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.397749 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.397770 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.397786 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:43Z","lastTransitionTime":"2026-02-14T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.499930 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.499962 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.499973 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.499989 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.500000 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:43Z","lastTransitionTime":"2026-02-14T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.602423 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.602460 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.602471 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.602487 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.602498 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:43Z","lastTransitionTime":"2026-02-14T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.704496 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.704533 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.704541 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.704554 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.704564 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:43Z","lastTransitionTime":"2026-02-14T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.806651 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.806686 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.806696 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.806711 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.806721 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:43Z","lastTransitionTime":"2026-02-14T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.836027 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.836070 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.836106 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:43 crc kubenswrapper[4904]: E0214 11:11:43.836155 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:43 crc kubenswrapper[4904]: E0214 11:11:43.836215 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:43 crc kubenswrapper[4904]: E0214 11:11:43.836343 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.851987 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 21:24:01.648184639 +0000 UTC Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.908972 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.909046 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.909058 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.909072 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:43 crc kubenswrapper[4904]: I0214 11:11:43.909081 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:43Z","lastTransitionTime":"2026-02-14T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.011134 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.011181 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.011193 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.011209 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.011222 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:44Z","lastTransitionTime":"2026-02-14T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.113293 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.113341 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.113356 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.113376 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.113391 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:44Z","lastTransitionTime":"2026-02-14T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.216584 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.216644 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.216666 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.216771 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.216792 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:44Z","lastTransitionTime":"2026-02-14T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.319890 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.319960 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.319985 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.320014 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.320037 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:44Z","lastTransitionTime":"2026-02-14T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.423013 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.423072 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.423088 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.423104 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.423114 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:44Z","lastTransitionTime":"2026-02-14T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.525714 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.525779 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.525802 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.525831 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.525891 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:44Z","lastTransitionTime":"2026-02-14T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.629154 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.629258 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.629274 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.629291 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.629333 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:44Z","lastTransitionTime":"2026-02-14T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.732252 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.732325 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.732350 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.732383 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.732412 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:44Z","lastTransitionTime":"2026-02-14T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.835154 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.835205 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.835226 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.835257 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.835280 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:44Z","lastTransitionTime":"2026-02-14T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.835489 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:44 crc kubenswrapper[4904]: E0214 11:11:44.835639 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.853128 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 05:45:18.877602788 +0000 UTC Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.938222 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.938265 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.938281 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.938303 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:44 crc kubenswrapper[4904]: I0214 11:11:44.938320 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:44Z","lastTransitionTime":"2026-02-14T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.040638 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.040678 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.040693 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.040711 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.040724 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:45Z","lastTransitionTime":"2026-02-14T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.142983 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.143018 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.143026 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.143037 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.143047 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:45Z","lastTransitionTime":"2026-02-14T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.245825 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.246927 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.246976 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.247031 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.247045 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:45Z","lastTransitionTime":"2026-02-14T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.349871 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.349924 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.349936 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.349953 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.349966 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:45Z","lastTransitionTime":"2026-02-14T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.454024 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.454063 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.454078 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.454097 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.454111 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:45Z","lastTransitionTime":"2026-02-14T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.556535 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.556573 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.556587 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.556606 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.556619 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:45Z","lastTransitionTime":"2026-02-14T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.658565 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.658600 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.658610 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.658625 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.658636 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:45Z","lastTransitionTime":"2026-02-14T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.768372 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.768429 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.768449 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.768476 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.768497 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:45Z","lastTransitionTime":"2026-02-14T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.835868 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.836030 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:45 crc kubenswrapper[4904]: E0214 11:11:45.836248 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.836271 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:45 crc kubenswrapper[4904]: E0214 11:11:45.836659 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:45 crc kubenswrapper[4904]: E0214 11:11:45.837172 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.853992 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 12:27:41.520548602 +0000 UTC Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.871119 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.871165 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.871176 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.871193 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.871206 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:45Z","lastTransitionTime":"2026-02-14T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.973450 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.973489 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.973503 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.973521 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:45 crc kubenswrapper[4904]: I0214 11:11:45.973533 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:45Z","lastTransitionTime":"2026-02-14T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.077114 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.077178 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.077191 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.077525 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.077543 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:46Z","lastTransitionTime":"2026-02-14T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.180743 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.180791 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.180800 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.180812 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.180821 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:46Z","lastTransitionTime":"2026-02-14T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.283605 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.283643 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.283655 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.283670 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.283681 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:46Z","lastTransitionTime":"2026-02-14T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.386058 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.386102 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.386112 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.386130 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.386140 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:46Z","lastTransitionTime":"2026-02-14T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.487749 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.487798 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.487811 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.487854 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.487867 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:46Z","lastTransitionTime":"2026-02-14T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.589861 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.589903 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.589918 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.589939 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.589953 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:46Z","lastTransitionTime":"2026-02-14T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.691913 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.691953 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.691965 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.691982 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.691994 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:46Z","lastTransitionTime":"2026-02-14T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.794392 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.794425 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.794433 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.794448 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.794457 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:46Z","lastTransitionTime":"2026-02-14T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.835913 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:46 crc kubenswrapper[4904]: E0214 11:11:46.836032 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.847585 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.847624 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.847633 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.847647 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.847658 4904 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-14T11:11:46Z","lastTransitionTime":"2026-02-14T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.854171 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 12:56:37.804662895 +0000 UTC Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.889966 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v"] Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.890606 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.893581 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.893746 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.893980 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 14 11:11:46 crc kubenswrapper[4904]: I0214 11:11:46.894045 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.091789 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d036a8d7-8be9-4c96-b496-a4d70f1c4f75-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ldr9v\" (UID: \"d036a8d7-8be9-4c96-b496-a4d70f1c4f75\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.091945 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d036a8d7-8be9-4c96-b496-a4d70f1c4f75-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ldr9v\" (UID: \"d036a8d7-8be9-4c96-b496-a4d70f1c4f75\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.091976 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d036a8d7-8be9-4c96-b496-a4d70f1c4f75-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ldr9v\" (UID: \"d036a8d7-8be9-4c96-b496-a4d70f1c4f75\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.092026 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d036a8d7-8be9-4c96-b496-a4d70f1c4f75-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ldr9v\" (UID: \"d036a8d7-8be9-4c96-b496-a4d70f1c4f75\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.092049 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d036a8d7-8be9-4c96-b496-a4d70f1c4f75-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ldr9v\" (UID: \"d036a8d7-8be9-4c96-b496-a4d70f1c4f75\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.193437 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d036a8d7-8be9-4c96-b496-a4d70f1c4f75-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ldr9v\" (UID: \"d036a8d7-8be9-4c96-b496-a4d70f1c4f75\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.193522 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d036a8d7-8be9-4c96-b496-a4d70f1c4f75-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ldr9v\" (UID: \"d036a8d7-8be9-4c96-b496-a4d70f1c4f75\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.193579 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d036a8d7-8be9-4c96-b496-a4d70f1c4f75-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ldr9v\" (UID: \"d036a8d7-8be9-4c96-b496-a4d70f1c4f75\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.193645 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d036a8d7-8be9-4c96-b496-a4d70f1c4f75-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ldr9v\" (UID: \"d036a8d7-8be9-4c96-b496-a4d70f1c4f75\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.193670 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d036a8d7-8be9-4c96-b496-a4d70f1c4f75-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ldr9v\" (UID: \"d036a8d7-8be9-4c96-b496-a4d70f1c4f75\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.194132 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d036a8d7-8be9-4c96-b496-a4d70f1c4f75-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ldr9v\" (UID: \"d036a8d7-8be9-4c96-b496-a4d70f1c4f75\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.194134 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d036a8d7-8be9-4c96-b496-a4d70f1c4f75-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ldr9v\" (UID: \"d036a8d7-8be9-4c96-b496-a4d70f1c4f75\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.194905 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d036a8d7-8be9-4c96-b496-a4d70f1c4f75-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ldr9v\" (UID: \"d036a8d7-8be9-4c96-b496-a4d70f1c4f75\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.202499 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d036a8d7-8be9-4c96-b496-a4d70f1c4f75-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ldr9v\" (UID: \"d036a8d7-8be9-4c96-b496-a4d70f1c4f75\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.214574 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d036a8d7-8be9-4c96-b496-a4d70f1c4f75-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ldr9v\" (UID: \"d036a8d7-8be9-4c96-b496-a4d70f1c4f75\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.508439 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.621619 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" event={"ID":"d036a8d7-8be9-4c96-b496-a4d70f1c4f75","Type":"ContainerStarted","Data":"792635e7f6e208e8ad59c3d5b36eed75a01ffc4cf1c314d230961f410f262b97"} Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.621656 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" event={"ID":"d036a8d7-8be9-4c96-b496-a4d70f1c4f75","Type":"ContainerStarted","Data":"e312b1b272c9a077ed777ad8184f60662d01e745b4c24144f29aa7c9d675033a"} Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.836335 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:47 crc kubenswrapper[4904]: E0214 11:11:47.836788 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.837059 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:47 crc kubenswrapper[4904]: E0214 11:11:47.837171 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.837061 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:47 crc kubenswrapper[4904]: E0214 11:11:47.837246 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.854334 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 10:55:07.957874065 +0000 UTC Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.854407 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 14 11:11:47 crc kubenswrapper[4904]: I0214 11:11:47.861856 4904 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 14 11:11:48 crc kubenswrapper[4904]: I0214 11:11:48.710027 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs\") pod \"network-metrics-daemon-fz5f2\" (UID: \"ac91dd23-e4e5-4073-af76-d760dfdd1adc\") " pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:48 crc kubenswrapper[4904]: E0214 11:11:48.710136 4904 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 14 11:11:48 crc kubenswrapper[4904]: E0214 11:11:48.710207 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs podName:ac91dd23-e4e5-4073-af76-d760dfdd1adc nodeName:}" failed. No retries permitted until 2026-02-14 11:12:52.710189026 +0000 UTC m=+163.522953687 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs") pod "network-metrics-daemon-fz5f2" (UID: "ac91dd23-e4e5-4073-af76-d760dfdd1adc") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 14 11:11:48 crc kubenswrapper[4904]: I0214 11:11:48.836194 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:48 crc kubenswrapper[4904]: E0214 11:11:48.836292 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:49 crc kubenswrapper[4904]: I0214 11:11:49.835871 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:49 crc kubenswrapper[4904]: I0214 11:11:49.835883 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:49 crc kubenswrapper[4904]: I0214 11:11:49.835934 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:49 crc kubenswrapper[4904]: E0214 11:11:49.836749 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:49 crc kubenswrapper[4904]: E0214 11:11:49.836847 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:49 crc kubenswrapper[4904]: E0214 11:11:49.836894 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:50 crc kubenswrapper[4904]: I0214 11:11:50.835427 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:50 crc kubenswrapper[4904]: E0214 11:11:50.835522 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:51 crc kubenswrapper[4904]: I0214 11:11:51.835528 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:51 crc kubenswrapper[4904]: I0214 11:11:51.835586 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:51 crc kubenswrapper[4904]: I0214 11:11:51.835574 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:51 crc kubenswrapper[4904]: E0214 11:11:51.835750 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:51 crc kubenswrapper[4904]: E0214 11:11:51.835850 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:51 crc kubenswrapper[4904]: E0214 11:11:51.835912 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:52 crc kubenswrapper[4904]: I0214 11:11:52.835471 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:52 crc kubenswrapper[4904]: E0214 11:11:52.835672 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:53 crc kubenswrapper[4904]: I0214 11:11:53.835738 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:53 crc kubenswrapper[4904]: E0214 11:11:53.835929 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:53 crc kubenswrapper[4904]: I0214 11:11:53.836185 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:53 crc kubenswrapper[4904]: I0214 11:11:53.835767 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:53 crc kubenswrapper[4904]: E0214 11:11:53.836279 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:53 crc kubenswrapper[4904]: E0214 11:11:53.837242 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:54 crc kubenswrapper[4904]: I0214 11:11:54.974004 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:54 crc kubenswrapper[4904]: I0214 11:11:54.974074 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:54 crc kubenswrapper[4904]: I0214 11:11:54.974110 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:54 crc kubenswrapper[4904]: E0214 11:11:54.974319 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:54 crc kubenswrapper[4904]: E0214 11:11:54.974997 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:54 crc kubenswrapper[4904]: E0214 11:11:54.975238 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:54 crc kubenswrapper[4904]: I0214 11:11:54.975901 4904 scope.go:117] "RemoveContainer" containerID="9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70" Feb 14 11:11:54 crc kubenswrapper[4904]: E0214 11:11:54.976216 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2gwlw_openshift-ovn-kubernetes(1ca5acc0-8b15-4052-883f-1987b7ca0067)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" Feb 14 11:11:55 crc kubenswrapper[4904]: I0214 11:11:55.835977 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:55 crc kubenswrapper[4904]: E0214 11:11:55.836159 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:56 crc kubenswrapper[4904]: I0214 11:11:56.835406 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:56 crc kubenswrapper[4904]: I0214 11:11:56.835685 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:56 crc kubenswrapper[4904]: E0214 11:11:56.835804 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:56 crc kubenswrapper[4904]: I0214 11:11:56.836017 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:56 crc kubenswrapper[4904]: E0214 11:11:56.836043 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:56 crc kubenswrapper[4904]: E0214 11:11:56.836120 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:57 crc kubenswrapper[4904]: I0214 11:11:57.835669 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:57 crc kubenswrapper[4904]: E0214 11:11:57.835779 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:11:58 crc kubenswrapper[4904]: I0214 11:11:58.835797 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:11:58 crc kubenswrapper[4904]: E0214 11:11:58.835932 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:11:58 crc kubenswrapper[4904]: I0214 11:11:58.836060 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:11:58 crc kubenswrapper[4904]: I0214 11:11:58.836127 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:11:58 crc kubenswrapper[4904]: E0214 11:11:58.836756 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:11:58 crc kubenswrapper[4904]: E0214 11:11:58.836492 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:11:59 crc kubenswrapper[4904]: I0214 11:11:59.835937 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:11:59 crc kubenswrapper[4904]: E0214 11:11:59.837817 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:12:00 crc kubenswrapper[4904]: I0214 11:12:00.835856 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:12:00 crc kubenswrapper[4904]: I0214 11:12:00.836098 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:12:00 crc kubenswrapper[4904]: E0214 11:12:00.836202 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:12:00 crc kubenswrapper[4904]: I0214 11:12:00.836117 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:12:00 crc kubenswrapper[4904]: E0214 11:12:00.836435 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:12:00 crc kubenswrapper[4904]: E0214 11:12:00.836489 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:12:01 crc kubenswrapper[4904]: I0214 11:12:01.836112 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:12:01 crc kubenswrapper[4904]: E0214 11:12:01.836292 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:12:02 crc kubenswrapper[4904]: I0214 11:12:02.836021 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:12:02 crc kubenswrapper[4904]: I0214 11:12:02.836094 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:12:02 crc kubenswrapper[4904]: E0214 11:12:02.836138 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:12:02 crc kubenswrapper[4904]: E0214 11:12:02.836291 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:12:02 crc kubenswrapper[4904]: I0214 11:12:02.836344 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:12:02 crc kubenswrapper[4904]: E0214 11:12:02.836397 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:12:03 crc kubenswrapper[4904]: I0214 11:12:03.835882 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:12:03 crc kubenswrapper[4904]: E0214 11:12:03.836005 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:12:04 crc kubenswrapper[4904]: I0214 11:12:04.836195 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:12:04 crc kubenswrapper[4904]: I0214 11:12:04.836195 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:12:04 crc kubenswrapper[4904]: I0214 11:12:04.836227 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:12:04 crc kubenswrapper[4904]: E0214 11:12:04.836422 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:12:04 crc kubenswrapper[4904]: E0214 11:12:04.836517 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:12:04 crc kubenswrapper[4904]: E0214 11:12:04.836662 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:12:05 crc kubenswrapper[4904]: I0214 11:12:05.685797 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jz5f5_0b01cc02-2e62-46e1-b07d-b3c0bff2b050/kube-multus/1.log" Feb 14 11:12:05 crc kubenswrapper[4904]: I0214 11:12:05.686691 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jz5f5_0b01cc02-2e62-46e1-b07d-b3c0bff2b050/kube-multus/0.log" Feb 14 11:12:05 crc kubenswrapper[4904]: I0214 11:12:05.686755 4904 generic.go:334] "Generic (PLEG): container finished" podID="0b01cc02-2e62-46e1-b07d-b3c0bff2b050" containerID="58ce2d3b3613f2f1c1ec36d61edb9bb614e5c1cdf37132359c08ef1f142d8bb0" exitCode=1 Feb 14 11:12:05 crc kubenswrapper[4904]: I0214 11:12:05.686793 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jz5f5" event={"ID":"0b01cc02-2e62-46e1-b07d-b3c0bff2b050","Type":"ContainerDied","Data":"58ce2d3b3613f2f1c1ec36d61edb9bb614e5c1cdf37132359c08ef1f142d8bb0"} Feb 14 11:12:05 crc kubenswrapper[4904]: I0214 11:12:05.686873 4904 scope.go:117] "RemoveContainer" containerID="9fb41c61d6bea1ad0d24dc40784b126f666eecb54ebbbaaec4c1af23d2538938" Feb 14 11:12:05 crc kubenswrapper[4904]: I0214 11:12:05.687190 4904 scope.go:117] "RemoveContainer" containerID="58ce2d3b3613f2f1c1ec36d61edb9bb614e5c1cdf37132359c08ef1f142d8bb0" Feb 14 11:12:05 crc kubenswrapper[4904]: E0214 11:12:05.687338 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-jz5f5_openshift-multus(0b01cc02-2e62-46e1-b07d-b3c0bff2b050)\"" pod="openshift-multus/multus-jz5f5" podUID="0b01cc02-2e62-46e1-b07d-b3c0bff2b050" Feb 14 11:12:05 crc kubenswrapper[4904]: I0214 11:12:05.705797 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ldr9v" podStartSLOduration=95.705766401 podStartE2EDuration="1m35.705766401s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:11:48.642240971 +0000 UTC m=+99.455005662" watchObservedRunningTime="2026-02-14 11:12:05.705766401 +0000 UTC m=+116.518531062" Feb 14 11:12:05 crc kubenswrapper[4904]: I0214 11:12:05.835430 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:12:05 crc kubenswrapper[4904]: E0214 11:12:05.835587 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:12:06 crc kubenswrapper[4904]: I0214 11:12:06.691421 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jz5f5_0b01cc02-2e62-46e1-b07d-b3c0bff2b050/kube-multus/1.log" Feb 14 11:12:06 crc kubenswrapper[4904]: I0214 11:12:06.835360 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:12:06 crc kubenswrapper[4904]: I0214 11:12:06.835372 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:12:06 crc kubenswrapper[4904]: E0214 11:12:06.835542 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:12:06 crc kubenswrapper[4904]: I0214 11:12:06.835390 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:12:06 crc kubenswrapper[4904]: E0214 11:12:06.835598 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:12:06 crc kubenswrapper[4904]: E0214 11:12:06.835678 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:12:07 crc kubenswrapper[4904]: I0214 11:12:07.836357 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:12:07 crc kubenswrapper[4904]: E0214 11:12:07.836553 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:12:08 crc kubenswrapper[4904]: I0214 11:12:08.836413 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:12:08 crc kubenswrapper[4904]: I0214 11:12:08.836481 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:12:08 crc kubenswrapper[4904]: E0214 11:12:08.836583 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:12:08 crc kubenswrapper[4904]: E0214 11:12:08.836782 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:12:08 crc kubenswrapper[4904]: I0214 11:12:08.836406 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:12:08 crc kubenswrapper[4904]: E0214 11:12:08.836945 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:12:08 crc kubenswrapper[4904]: I0214 11:12:08.838587 4904 scope.go:117] "RemoveContainer" containerID="9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70" Feb 14 11:12:09 crc kubenswrapper[4904]: I0214 11:12:09.701061 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovnkube-controller/3.log" Feb 14 11:12:09 crc kubenswrapper[4904]: I0214 11:12:09.703314 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerStarted","Data":"8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6"} Feb 14 11:12:09 crc kubenswrapper[4904]: I0214 11:12:09.703791 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:12:09 crc kubenswrapper[4904]: I0214 11:12:09.731565 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-fz5f2"] Feb 14 11:12:09 crc kubenswrapper[4904]: I0214 11:12:09.731683 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:12:09 crc kubenswrapper[4904]: E0214 11:12:09.731786 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:12:09 crc kubenswrapper[4904]: I0214 11:12:09.740977 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" podStartSLOduration=99.740956464 podStartE2EDuration="1m39.740956464s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:09.738133699 +0000 UTC m=+120.550898350" watchObservedRunningTime="2026-02-14 11:12:09.740956464 +0000 UTC m=+120.553721135" Feb 14 11:12:09 crc kubenswrapper[4904]: E0214 11:12:09.830081 4904 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 14 11:12:09 crc kubenswrapper[4904]: I0214 11:12:09.835362 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:12:09 crc kubenswrapper[4904]: E0214 11:12:09.836909 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:12:09 crc kubenswrapper[4904]: E0214 11:12:09.980648 4904 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 14 11:12:10 crc kubenswrapper[4904]: I0214 11:12:10.836057 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:12:10 crc kubenswrapper[4904]: I0214 11:12:10.836138 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:12:10 crc kubenswrapper[4904]: E0214 11:12:10.836464 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:12:10 crc kubenswrapper[4904]: E0214 11:12:10.836594 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:12:11 crc kubenswrapper[4904]: I0214 11:12:11.836312 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:12:11 crc kubenswrapper[4904]: I0214 11:12:11.836328 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:12:11 crc kubenswrapper[4904]: E0214 11:12:11.836470 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:12:11 crc kubenswrapper[4904]: E0214 11:12:11.836691 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:12:12 crc kubenswrapper[4904]: I0214 11:12:12.835543 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:12:12 crc kubenswrapper[4904]: I0214 11:12:12.835637 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:12:12 crc kubenswrapper[4904]: E0214 11:12:12.835748 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:12:12 crc kubenswrapper[4904]: E0214 11:12:12.835947 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:12:13 crc kubenswrapper[4904]: I0214 11:12:13.835390 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:12:13 crc kubenswrapper[4904]: E0214 11:12:13.835576 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:12:13 crc kubenswrapper[4904]: I0214 11:12:13.835922 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:12:13 crc kubenswrapper[4904]: E0214 11:12:13.836021 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:12:14 crc kubenswrapper[4904]: I0214 11:12:14.835744 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:12:14 crc kubenswrapper[4904]: I0214 11:12:14.835798 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:12:14 crc kubenswrapper[4904]: E0214 11:12:14.835911 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:12:14 crc kubenswrapper[4904]: E0214 11:12:14.836017 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:12:14 crc kubenswrapper[4904]: E0214 11:12:14.982379 4904 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 14 11:12:15 crc kubenswrapper[4904]: I0214 11:12:15.835667 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:12:15 crc kubenswrapper[4904]: I0214 11:12:15.835706 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:12:15 crc kubenswrapper[4904]: E0214 11:12:15.835963 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:12:15 crc kubenswrapper[4904]: E0214 11:12:15.836089 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:12:16 crc kubenswrapper[4904]: I0214 11:12:16.835384 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:12:16 crc kubenswrapper[4904]: I0214 11:12:16.835391 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:12:16 crc kubenswrapper[4904]: E0214 11:12:16.835520 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:12:16 crc kubenswrapper[4904]: E0214 11:12:16.836424 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:12:17 crc kubenswrapper[4904]: I0214 11:12:17.835930 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:12:17 crc kubenswrapper[4904]: I0214 11:12:17.835930 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:12:17 crc kubenswrapper[4904]: E0214 11:12:17.836090 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:12:17 crc kubenswrapper[4904]: E0214 11:12:17.836186 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:12:18 crc kubenswrapper[4904]: I0214 11:12:18.835777 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:12:18 crc kubenswrapper[4904]: I0214 11:12:18.835915 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:12:18 crc kubenswrapper[4904]: E0214 11:12:18.836007 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:12:18 crc kubenswrapper[4904]: E0214 11:12:18.836314 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:12:18 crc kubenswrapper[4904]: I0214 11:12:18.836581 4904 scope.go:117] "RemoveContainer" containerID="58ce2d3b3613f2f1c1ec36d61edb9bb614e5c1cdf37132359c08ef1f142d8bb0" Feb 14 11:12:19 crc kubenswrapper[4904]: I0214 11:12:19.739128 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jz5f5_0b01cc02-2e62-46e1-b07d-b3c0bff2b050/kube-multus/1.log" Feb 14 11:12:19 crc kubenswrapper[4904]: I0214 11:12:19.739447 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jz5f5" event={"ID":"0b01cc02-2e62-46e1-b07d-b3c0bff2b050","Type":"ContainerStarted","Data":"a746e87d6bbe5102b3055a0f666d693b83f8f709786e5b8ab22d2cbaf9a6b7ee"} Feb 14 11:12:19 crc kubenswrapper[4904]: I0214 11:12:19.836623 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:12:19 crc kubenswrapper[4904]: E0214 11:12:19.838924 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:12:19 crc kubenswrapper[4904]: I0214 11:12:19.840728 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:12:19 crc kubenswrapper[4904]: E0214 11:12:19.840948 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:12:19 crc kubenswrapper[4904]: E0214 11:12:19.983565 4904 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 14 11:12:20 crc kubenswrapper[4904]: I0214 11:12:20.835866 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:12:20 crc kubenswrapper[4904]: I0214 11:12:20.836065 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:12:20 crc kubenswrapper[4904]: E0214 11:12:20.836135 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:12:20 crc kubenswrapper[4904]: E0214 11:12:20.836398 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:12:21 crc kubenswrapper[4904]: I0214 11:12:21.836389 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:12:21 crc kubenswrapper[4904]: E0214 11:12:21.836641 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:12:21 crc kubenswrapper[4904]: I0214 11:12:21.837065 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:12:21 crc kubenswrapper[4904]: E0214 11:12:21.837250 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:12:22 crc kubenswrapper[4904]: I0214 11:12:22.835957 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:12:22 crc kubenswrapper[4904]: E0214 11:12:22.836073 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:12:22 crc kubenswrapper[4904]: I0214 11:12:22.835957 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:12:22 crc kubenswrapper[4904]: E0214 11:12:22.836197 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:12:23 crc kubenswrapper[4904]: I0214 11:12:23.835398 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:12:23 crc kubenswrapper[4904]: E0214 11:12:23.835515 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fz5f2" podUID="ac91dd23-e4e5-4073-af76-d760dfdd1adc" Feb 14 11:12:23 crc kubenswrapper[4904]: I0214 11:12:23.835666 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:12:23 crc kubenswrapper[4904]: E0214 11:12:23.835714 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 14 11:12:24 crc kubenswrapper[4904]: I0214 11:12:24.835289 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:12:24 crc kubenswrapper[4904]: E0214 11:12:24.835414 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 14 11:12:24 crc kubenswrapper[4904]: I0214 11:12:24.835471 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:12:24 crc kubenswrapper[4904]: E0214 11:12:24.835677 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 14 11:12:25 crc kubenswrapper[4904]: I0214 11:12:25.836133 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:12:25 crc kubenswrapper[4904]: I0214 11:12:25.836133 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:12:25 crc kubenswrapper[4904]: I0214 11:12:25.839995 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 14 11:12:25 crc kubenswrapper[4904]: I0214 11:12:25.840271 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 14 11:12:25 crc kubenswrapper[4904]: I0214 11:12:25.840313 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 14 11:12:25 crc kubenswrapper[4904]: I0214 11:12:25.840399 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 14 11:12:26 crc kubenswrapper[4904]: I0214 11:12:26.836099 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:12:26 crc kubenswrapper[4904]: I0214 11:12:26.836181 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:12:26 crc kubenswrapper[4904]: I0214 11:12:26.838167 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 14 11:12:26 crc kubenswrapper[4904]: I0214 11:12:26.839835 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.594265 4904 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.622757 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-t8ggt"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.623455 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.626793 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-hgggh"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.627313 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-hgggh" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.627824 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.629002 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.629162 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.629310 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.637022 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.638076 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.638223 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.638343 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jq824"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.638661 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.639091 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.640591 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.640706 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.640799 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.641037 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.641152 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.641328 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.641449 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.643800 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.644057 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.644194 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.644891 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.644995 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.647466 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.647542 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5lwz5"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.647712 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.647795 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g25mq"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.648089 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-f95s6"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.648177 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.648189 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.648303 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-f95s6" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.648370 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.648580 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.648887 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-etcd-client\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.648909 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/800acf17-404a-4dc3-a25f-66e779d292fd-config\") pod \"machine-api-operator-5694c8668f-hgggh\" (UID: \"800acf17-404a-4dc3-a25f-66e779d292fd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hgggh" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.648932 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-image-import-ca\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.648949 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-trusted-ca-bundle\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.648969 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-node-pullsecrets\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.648984 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-etcd-serving-ca\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.649001 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/800acf17-404a-4dc3-a25f-66e779d292fd-images\") pod \"machine-api-operator-5694c8668f-hgggh\" (UID: \"800acf17-404a-4dc3-a25f-66e779d292fd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hgggh" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.649015 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n56s2\" (UniqueName: \"kubernetes.io/projected/800acf17-404a-4dc3-a25f-66e779d292fd-kube-api-access-n56s2\") pod \"machine-api-operator-5694c8668f-hgggh\" (UID: \"800acf17-404a-4dc3-a25f-66e779d292fd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hgggh" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.649028 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-serving-cert\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.655820 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/800acf17-404a-4dc3-a25f-66e779d292fd-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-hgggh\" (UID: \"800acf17-404a-4dc3-a25f-66e779d292fd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hgggh" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.655896 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-config\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.655925 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-audit-dir\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.655964 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfr2m\" (UniqueName: \"kubernetes.io/projected/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-kube-api-access-qfr2m\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.655987 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-encryption-config\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.656010 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-audit\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.660306 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-bzps5"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.661640 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-bzps5" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.662368 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g25mq" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.667431 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.667720 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.668111 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.668160 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.668606 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.669078 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.670185 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.670805 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.671280 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.672246 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.672989 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.673078 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.673127 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.673199 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.673300 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.673415 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.673422 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.673535 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.673576 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.673539 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.673658 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qsppc"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.674174 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qsppc" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.674705 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7t272"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.674719 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.675597 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7t272" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.677228 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.677701 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.680699 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.682016 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.682816 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.685222 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.691570 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.692180 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.692694 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.692810 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.693137 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-pcc6z"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.693160 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.693786 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pcc6z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.694197 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.694798 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.695923 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-d948p"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.696596 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.697556 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.697623 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.697625 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.697702 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.697659 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.697824 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.697920 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.697980 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.698082 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.698129 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.700012 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.700267 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-d9d4k"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.700526 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ql7rx"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.700782 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fqtd8"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.701201 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fqtd8" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.701528 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-d948p" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.701749 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.701995 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ql7rx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.712663 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.712692 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.712860 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.712946 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.713022 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.713051 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.713082 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.713155 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.713178 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.713194 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.713249 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.713298 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.713372 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.713432 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.713084 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.713493 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.713406 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.712703 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.718230 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.713495 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.718657 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.718733 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.719147 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.719535 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.719926 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.720324 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.720450 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.720916 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.721168 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.721451 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.721586 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.721741 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.722017 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.723990 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.724186 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.724566 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.725067 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.713414 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.725274 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.725821 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.728023 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.737955 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7h7pz"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.738440 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.738717 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.738928 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-7h7pz" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.739556 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-x4nlj"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.739929 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-qjrkr"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.740309 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qjrkr" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.740898 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-49jn7"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.741098 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.741518 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-x4nlj" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.742269 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.742587 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.742691 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-2gsdv"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.742772 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-49jn7" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.743418 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2gsdv" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.743713 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c5fmb"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.744176 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c5fmb" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.746290 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4n756"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.746653 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4n756" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.748069 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-qh6sr"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.748428 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.748805 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjcqb"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.749388 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjcqb" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.749875 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zbh4h"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.750227 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.752350 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.752687 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bfn42"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.753035 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.753292 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.753782 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hfw59"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.754498 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hfw59" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.756328 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-7c6cp"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.756888 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.757268 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.757586 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.757770 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9pzkg"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758231 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758432 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/821984e5-b740-4332-9c06-1ccefda9a9fb-encryption-config\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758472 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3253b15-b01e-4c6b-91df-578ddb39b234-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jq824\" (UID: \"e3253b15-b01e-4c6b-91df-578ddb39b234\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758497 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/345bb34c-a9b2-4e88-94ce-3187b8933057-serving-cert\") pod \"openshift-config-operator-7777fb866f-4kpjx\" (UID: \"345bb34c-a9b2-4e88-94ce-3187b8933057\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758523 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/821984e5-b740-4332-9c06-1ccefda9a9fb-audit-policies\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758577 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b42c5242-55b6-43aa-a9bf-e368d446c3a5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qsppc\" (UID: \"b42c5242-55b6-43aa-a9bf-e368d446c3a5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qsppc" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758605 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/de09ed81-3849-460e-93f2-fb0f5d72a4b9-tmpfs\") pod \"packageserver-d55dfcdfc-5vpqp\" (UID: \"de09ed81-3849-460e-93f2-fb0f5d72a4b9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758631 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-etcd-client\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758652 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b42c5242-55b6-43aa-a9bf-e368d446c3a5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qsppc\" (UID: \"b42c5242-55b6-43aa-a9bf-e368d446c3a5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qsppc" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758676 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cszmb\" (UniqueName: \"kubernetes.io/projected/2201ed6d-9862-43d5-98ed-949539d6d0ae-kube-api-access-cszmb\") pod \"kube-storage-version-migrator-operator-b67b599dd-ql7rx\" (UID: \"2201ed6d-9862-43d5-98ed-949539d6d0ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ql7rx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758707 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/800acf17-404a-4dc3-a25f-66e779d292fd-config\") pod \"machine-api-operator-5694c8668f-hgggh\" (UID: \"800acf17-404a-4dc3-a25f-66e779d292fd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hgggh" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758729 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a7f9f078-ff94-4adc-be45-927915d6e55f-etcd-service-ca\") pod \"etcd-operator-b45778765-d9d4k\" (UID: \"a7f9f078-ff94-4adc-be45-927915d6e55f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758753 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ddd4a23a-8856-480c-91d6-829f766120a9-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-vdl4w\" (UID: \"ddd4a23a-8856-480c-91d6-829f766120a9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758775 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/66d09685-dfb3-4344-8cb2-f4e820e05bac-images\") pod \"machine-config-operator-74547568cd-6rh7z\" (UID: \"66d09685-dfb3-4344-8cb2-f4e820e05bac\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758796 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/821984e5-b740-4332-9c06-1ccefda9a9fb-etcd-client\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758815 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/b9b87214-6f46-4d58-b2fe-5618cfe4e0f0-machine-approver-tls\") pod \"machine-approver-56656f9798-pcc6z\" (UID: \"b9b87214-6f46-4d58-b2fe-5618cfe4e0f0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pcc6z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758866 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-image-import-ca\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758896 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-trusted-ca-bundle\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758919 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mpwg\" (UniqueName: \"kubernetes.io/projected/e3253b15-b01e-4c6b-91df-578ddb39b234-kube-api-access-8mpwg\") pod \"authentication-operator-69f744f599-jq824\" (UID: \"e3253b15-b01e-4c6b-91df-578ddb39b234\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758947 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/de09ed81-3849-460e-93f2-fb0f5d72a4b9-apiservice-cert\") pod \"packageserver-d55dfcdfc-5vpqp\" (UID: \"de09ed81-3849-460e-93f2-fb0f5d72a4b9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758968 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a7f9f078-ff94-4adc-be45-927915d6e55f-etcd-ca\") pod \"etcd-operator-b45778765-d9d4k\" (UID: \"a7f9f078-ff94-4adc-be45-927915d6e55f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.758990 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-node-pullsecrets\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.759009 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a7f9f078-ff94-4adc-be45-927915d6e55f-etcd-client\") pod \"etcd-operator-b45778765-d9d4k\" (UID: \"a7f9f078-ff94-4adc-be45-927915d6e55f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.759032 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzmbq\" (UniqueName: \"kubernetes.io/projected/fdb5ad89-a263-4f1f-bbe5-42bd5b91ea25-kube-api-access-vzmbq\") pod \"cluster-samples-operator-665b6dd947-g25mq\" (UID: \"fdb5ad89-a263-4f1f-bbe5-42bd5b91ea25\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g25mq" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.759056 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-etcd-serving-ca\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.759076 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b89659e-58f4-4c11-9839-2be0bdcd6e5e-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5pmk4\" (UID: \"8b89659e-58f4-4c11-9839-2be0bdcd6e5e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.759095 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c2c0ca2-4182-4c24-8710-c3b632e768f9-config\") pod \"console-operator-58897d9998-f95s6\" (UID: \"7c2c0ca2-4182-4c24-8710-c3b632e768f9\") " pod="openshift-console-operator/console-operator-58897d9998-f95s6" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.759114 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/821984e5-b740-4332-9c06-1ccefda9a9fb-audit-dir\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.759137 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpfx4\" (UniqueName: \"kubernetes.io/projected/5a537ec4-ab86-4bc9-81d2-3870472b767a-kube-api-access-dpfx4\") pod \"controller-manager-879f6c89f-5lwz5\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.759168 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/821984e5-b740-4332-9c06-1ccefda9a9fb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.759193 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq4bg\" (UniqueName: \"kubernetes.io/projected/8b89659e-58f4-4c11-9839-2be0bdcd6e5e-kube-api-access-nq4bg\") pod \"ingress-operator-5b745b69d9-5pmk4\" (UID: \"8b89659e-58f4-4c11-9839-2be0bdcd6e5e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.759215 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/30e3e22f-c17f-4efa-b5cd-33406a09def1-signing-key\") pod \"service-ca-9c57cc56f-d948p\" (UID: \"30e3e22f-c17f-4efa-b5cd-33406a09def1\") " pod="openshift-service-ca/service-ca-9c57cc56f-d948p" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.759239 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/800acf17-404a-4dc3-a25f-66e779d292fd-images\") pod \"machine-api-operator-5694c8668f-hgggh\" (UID: \"800acf17-404a-4dc3-a25f-66e779d292fd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hgggh" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.759260 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n56s2\" (UniqueName: \"kubernetes.io/projected/800acf17-404a-4dc3-a25f-66e779d292fd-kube-api-access-n56s2\") pod \"machine-api-operator-5694c8668f-hgggh\" (UID: \"800acf17-404a-4dc3-a25f-66e779d292fd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hgggh" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.759284 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-serving-cert\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.759307 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/de09ed81-3849-460e-93f2-fb0f5d72a4b9-webhook-cert\") pod \"packageserver-d55dfcdfc-5vpqp\" (UID: \"de09ed81-3849-460e-93f2-fb0f5d72a4b9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.760178 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-node-pullsecrets\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.760521 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/800acf17-404a-4dc3-a25f-66e779d292fd-images\") pod \"machine-api-operator-5694c8668f-hgggh\" (UID: \"800acf17-404a-4dc3-a25f-66e779d292fd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hgggh" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.760930 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/800acf17-404a-4dc3-a25f-66e779d292fd-config\") pod \"machine-api-operator-5694c8668f-hgggh\" (UID: \"800acf17-404a-4dc3-a25f-66e779d292fd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hgggh" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.760990 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-etcd-serving-ca\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.761066 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-image-import-ca\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.761052 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/345bb34c-a9b2-4e88-94ce-3187b8933057-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4kpjx\" (UID: \"345bb34c-a9b2-4e88-94ce-3187b8933057\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.761223 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/30e3e22f-c17f-4efa-b5cd-33406a09def1-signing-cabundle\") pod \"service-ca-9c57cc56f-d948p\" (UID: \"30e3e22f-c17f-4efa-b5cd-33406a09def1\") " pod="openshift-service-ca/service-ca-9c57cc56f-d948p" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.761295 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/967509da-8c27-49be-aa96-3a7129937ddd-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7t272\" (UID: \"967509da-8c27-49be-aa96-3a7129937ddd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7t272" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.761373 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/967509da-8c27-49be-aa96-3a7129937ddd-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7t272\" (UID: \"967509da-8c27-49be-aa96-3a7129937ddd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7t272" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.761457 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/800acf17-404a-4dc3-a25f-66e779d292fd-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-hgggh\" (UID: \"800acf17-404a-4dc3-a25f-66e779d292fd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hgggh" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.761713 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-trusted-ca-bundle\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.761761 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b89659e-58f4-4c11-9839-2be0bdcd6e5e-trusted-ca\") pod \"ingress-operator-5b745b69d9-5pmk4\" (UID: \"8b89659e-58f4-4c11-9839-2be0bdcd6e5e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.761794 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ddd4a23a-8856-480c-91d6-829f766120a9-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-vdl4w\" (UID: \"ddd4a23a-8856-480c-91d6-829f766120a9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.761857 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-config\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.761883 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdv8b\" (UniqueName: \"kubernetes.io/projected/b9b87214-6f46-4d58-b2fe-5618cfe4e0f0-kube-api-access-jdv8b\") pod \"machine-approver-56656f9798-pcc6z\" (UID: \"b9b87214-6f46-4d58-b2fe-5618cfe4e0f0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pcc6z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.761909 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsk2p\" (UniqueName: \"kubernetes.io/projected/ddd4a23a-8856-480c-91d6-829f766120a9-kube-api-access-wsk2p\") pod \"cluster-image-registry-operator-dc59b4c8b-vdl4w\" (UID: \"ddd4a23a-8856-480c-91d6-829f766120a9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.761934 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/11f24e26-e608-47a1-b8b1-8ac0159c1c5a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-fqtd8\" (UID: \"11f24e26-e608-47a1-b8b1-8ac0159c1c5a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fqtd8" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.761956 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2201ed6d-9862-43d5-98ed-949539d6d0ae-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ql7rx\" (UID: \"2201ed6d-9862-43d5-98ed-949539d6d0ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ql7rx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.761979 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-audit-dir\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.761998 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8b89659e-58f4-4c11-9839-2be0bdcd6e5e-metrics-tls\") pod \"ingress-operator-5b745b69d9-5pmk4\" (UID: \"8b89659e-58f4-4c11-9839-2be0bdcd6e5e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762023 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfr2m\" (UniqueName: \"kubernetes.io/projected/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-kube-api-access-qfr2m\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762038 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3253b15-b01e-4c6b-91df-578ddb39b234-service-ca-bundle\") pod \"authentication-operator-69f744f599-jq824\" (UID: \"e3253b15-b01e-4c6b-91df-578ddb39b234\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762053 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a537ec4-ab86-4bc9-81d2-3870472b767a-serving-cert\") pod \"controller-manager-879f6c89f-5lwz5\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762071 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-encryption-config\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762087 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5a537ec4-ab86-4bc9-81d2-3870472b767a-client-ca\") pod \"controller-manager-879f6c89f-5lwz5\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762102 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7f9f078-ff94-4adc-be45-927915d6e55f-config\") pod \"etcd-operator-b45778765-d9d4k\" (UID: \"a7f9f078-ff94-4adc-be45-927915d6e55f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762117 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3253b15-b01e-4c6b-91df-578ddb39b234-config\") pod \"authentication-operator-69f744f599-jq824\" (UID: \"e3253b15-b01e-4c6b-91df-578ddb39b234\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762132 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5a537ec4-ab86-4bc9-81d2-3870472b767a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-5lwz5\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762146 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b9b87214-6f46-4d58-b2fe-5618cfe4e0f0-auth-proxy-config\") pod \"machine-approver-56656f9798-pcc6z\" (UID: \"b9b87214-6f46-4d58-b2fe-5618cfe4e0f0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pcc6z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762161 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7vch\" (UniqueName: \"kubernetes.io/projected/345bb34c-a9b2-4e88-94ce-3187b8933057-kube-api-access-c7vch\") pod \"openshift-config-operator-7777fb866f-4kpjx\" (UID: \"345bb34c-a9b2-4e88-94ce-3187b8933057\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762185 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-audit\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762201 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9b87214-6f46-4d58-b2fe-5618cfe4e0f0-config\") pod \"machine-approver-56656f9798-pcc6z\" (UID: \"b9b87214-6f46-4d58-b2fe-5618cfe4e0f0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pcc6z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762216 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcq6t\" (UniqueName: \"kubernetes.io/projected/66d09685-dfb3-4344-8cb2-f4e820e05bac-kube-api-access-pcq6t\") pod \"machine-config-operator-74547568cd-6rh7z\" (UID: \"66d09685-dfb3-4344-8cb2-f4e820e05bac\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762232 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sp79\" (UniqueName: \"kubernetes.io/projected/11f24e26-e608-47a1-b8b1-8ac0159c1c5a-kube-api-access-2sp79\") pod \"package-server-manager-789f6589d5-fqtd8\" (UID: \"11f24e26-e608-47a1-b8b1-8ac0159c1c5a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fqtd8" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762249 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a537ec4-ab86-4bc9-81d2-3870472b767a-config\") pod \"controller-manager-879f6c89f-5lwz5\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762263 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b42c5242-55b6-43aa-a9bf-e368d446c3a5-config\") pod \"kube-controller-manager-operator-78b949d7b-qsppc\" (UID: \"b42c5242-55b6-43aa-a9bf-e368d446c3a5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qsppc" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762278 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/821984e5-b740-4332-9c06-1ccefda9a9fb-serving-cert\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762293 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/821984e5-b740-4332-9c06-1ccefda9a9fb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762307 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vjcd\" (UniqueName: \"kubernetes.io/projected/a7f9f078-ff94-4adc-be45-927915d6e55f-kube-api-access-6vjcd\") pod \"etcd-operator-b45778765-d9d4k\" (UID: \"a7f9f078-ff94-4adc-be45-927915d6e55f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762322 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/66d09685-dfb3-4344-8cb2-f4e820e05bac-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6rh7z\" (UID: \"66d09685-dfb3-4344-8cb2-f4e820e05bac\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762338 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x5dn\" (UniqueName: \"kubernetes.io/projected/30e3e22f-c17f-4efa-b5cd-33406a09def1-kube-api-access-7x5dn\") pod \"service-ca-9c57cc56f-d948p\" (UID: \"30e3e22f-c17f-4efa-b5cd-33406a09def1\") " pod="openshift-service-ca/service-ca-9c57cc56f-d948p" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762353 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c2c0ca2-4182-4c24-8710-c3b632e768f9-serving-cert\") pod \"console-operator-58897d9998-f95s6\" (UID: \"7c2c0ca2-4182-4c24-8710-c3b632e768f9\") " pod="openshift-console-operator/console-operator-58897d9998-f95s6" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762367 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffpfg\" (UniqueName: \"kubernetes.io/projected/3cb9bb68-546e-4e52-9a15-5aeee129a5b4-kube-api-access-ffpfg\") pod \"downloads-7954f5f757-bzps5\" (UID: \"3cb9bb68-546e-4e52-9a15-5aeee129a5b4\") " pod="openshift-console/downloads-7954f5f757-bzps5" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762382 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7f9f078-ff94-4adc-be45-927915d6e55f-serving-cert\") pod \"etcd-operator-b45778765-d9d4k\" (UID: \"a7f9f078-ff94-4adc-be45-927915d6e55f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762397 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/66d09685-dfb3-4344-8cb2-f4e820e05bac-proxy-tls\") pod \"machine-config-operator-74547568cd-6rh7z\" (UID: \"66d09685-dfb3-4344-8cb2-f4e820e05bac\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762412 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3253b15-b01e-4c6b-91df-578ddb39b234-serving-cert\") pod \"authentication-operator-69f744f599-jq824\" (UID: \"e3253b15-b01e-4c6b-91df-578ddb39b234\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762425 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66jvb\" (UniqueName: \"kubernetes.io/projected/de09ed81-3849-460e-93f2-fb0f5d72a4b9-kube-api-access-66jvb\") pod \"packageserver-d55dfcdfc-5vpqp\" (UID: \"de09ed81-3849-460e-93f2-fb0f5d72a4b9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762441 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/fdb5ad89-a263-4f1f-bbe5-42bd5b91ea25-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-g25mq\" (UID: \"fdb5ad89-a263-4f1f-bbe5-42bd5b91ea25\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g25mq" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762457 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2201ed6d-9862-43d5-98ed-949539d6d0ae-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ql7rx\" (UID: \"2201ed6d-9862-43d5-98ed-949539d6d0ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ql7rx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762481 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7c2c0ca2-4182-4c24-8710-c3b632e768f9-trusted-ca\") pod \"console-operator-58897d9998-f95s6\" (UID: \"7c2c0ca2-4182-4c24-8710-c3b632e768f9\") " pod="openshift-console-operator/console-operator-58897d9998-f95s6" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762496 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q99f\" (UniqueName: \"kubernetes.io/projected/7c2c0ca2-4182-4c24-8710-c3b632e768f9-kube-api-access-7q99f\") pod \"console-operator-58897d9998-f95s6\" (UID: \"7c2c0ca2-4182-4c24-8710-c3b632e768f9\") " pod="openshift-console-operator/console-operator-58897d9998-f95s6" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762510 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5k6l\" (UniqueName: \"kubernetes.io/projected/967509da-8c27-49be-aa96-3a7129937ddd-kube-api-access-m5k6l\") pod \"openshift-apiserver-operator-796bbdcf4f-7t272\" (UID: \"967509da-8c27-49be-aa96-3a7129937ddd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7t272" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762527 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6q44\" (UniqueName: \"kubernetes.io/projected/821984e5-b740-4332-9c06-1ccefda9a9fb-kube-api-access-v6q44\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.762540 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ddd4a23a-8856-480c-91d6-829f766120a9-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-vdl4w\" (UID: \"ddd4a23a-8856-480c-91d6-829f766120a9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.763017 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-config\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.763076 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-audit-dir\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.763407 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.763719 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.764813 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-audit\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.771789 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.775163 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-serving-cert\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.782993 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.783716 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-etcd-client\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.784210 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/800acf17-404a-4dc3-a25f-66e779d292fd-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-hgggh\" (UID: \"800acf17-404a-4dc3-a25f-66e779d292fd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hgggh" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.791185 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.793247 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-encryption-config\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.798942 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.798989 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5lwz5"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.801611 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-d948p"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.806447 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.810063 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jq824"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.814374 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-hgggh"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.821101 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.822350 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.823294 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7h7pz"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.832531 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7t272"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.832794 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-d9d4k"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.846208 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872213 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872253 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8b89659e-58f4-4c11-9839-2be0bdcd6e5e-metrics-tls\") pod \"ingress-operator-5b745b69d9-5pmk4\" (UID: \"8b89659e-58f4-4c11-9839-2be0bdcd6e5e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872280 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2201ed6d-9862-43d5-98ed-949539d6d0ae-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ql7rx\" (UID: \"2201ed6d-9862-43d5-98ed-949539d6d0ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ql7rx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872309 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86a63698-caa8-4173-8f86-1bb56087babb-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-49jn7\" (UID: \"86a63698-caa8-4173-8f86-1bb56087babb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-49jn7" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872327 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86a63698-caa8-4173-8f86-1bb56087babb-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-49jn7\" (UID: \"86a63698-caa8-4173-8f86-1bb56087babb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-49jn7" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872345 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872380 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3253b15-b01e-4c6b-91df-578ddb39b234-service-ca-bundle\") pod \"authentication-operator-69f744f599-jq824\" (UID: \"e3253b15-b01e-4c6b-91df-578ddb39b234\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872399 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a537ec4-ab86-4bc9-81d2-3870472b767a-serving-cert\") pod \"controller-manager-879f6c89f-5lwz5\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872416 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7f9f078-ff94-4adc-be45-927915d6e55f-config\") pod \"etcd-operator-b45778765-d9d4k\" (UID: \"a7f9f078-ff94-4adc-be45-927915d6e55f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872433 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5a537ec4-ab86-4bc9-81d2-3870472b767a-client-ca\") pod \"controller-manager-879f6c89f-5lwz5\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872449 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3253b15-b01e-4c6b-91df-578ddb39b234-config\") pod \"authentication-operator-69f744f599-jq824\" (UID: \"e3253b15-b01e-4c6b-91df-578ddb39b234\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872466 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5a537ec4-ab86-4bc9-81d2-3870472b767a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-5lwz5\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872488 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b9b87214-6f46-4d58-b2fe-5618cfe4e0f0-auth-proxy-config\") pod \"machine-approver-56656f9798-pcc6z\" (UID: \"b9b87214-6f46-4d58-b2fe-5618cfe4e0f0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pcc6z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872510 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7vch\" (UniqueName: \"kubernetes.io/projected/345bb34c-a9b2-4e88-94ce-3187b8933057-kube-api-access-c7vch\") pod \"openshift-config-operator-7777fb866f-4kpjx\" (UID: \"345bb34c-a9b2-4e88-94ce-3187b8933057\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872537 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9b87214-6f46-4d58-b2fe-5618cfe4e0f0-config\") pod \"machine-approver-56656f9798-pcc6z\" (UID: \"b9b87214-6f46-4d58-b2fe-5618cfe4e0f0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pcc6z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872558 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcq6t\" (UniqueName: \"kubernetes.io/projected/66d09685-dfb3-4344-8cb2-f4e820e05bac-kube-api-access-pcq6t\") pod \"machine-config-operator-74547568cd-6rh7z\" (UID: \"66d09685-dfb3-4344-8cb2-f4e820e05bac\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872578 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sp79\" (UniqueName: \"kubernetes.io/projected/11f24e26-e608-47a1-b8b1-8ac0159c1c5a-kube-api-access-2sp79\") pod \"package-server-manager-789f6589d5-fqtd8\" (UID: \"11f24e26-e608-47a1-b8b1-8ac0159c1c5a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fqtd8" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872598 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b42c5242-55b6-43aa-a9bf-e368d446c3a5-config\") pod \"kube-controller-manager-operator-78b949d7b-qsppc\" (UID: \"b42c5242-55b6-43aa-a9bf-e368d446c3a5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qsppc" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872618 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a537ec4-ab86-4bc9-81d2-3870472b767a-config\") pod \"controller-manager-879f6c89f-5lwz5\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872633 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vjcd\" (UniqueName: \"kubernetes.io/projected/a7f9f078-ff94-4adc-be45-927915d6e55f-kube-api-access-6vjcd\") pod \"etcd-operator-b45778765-d9d4k\" (UID: \"a7f9f078-ff94-4adc-be45-927915d6e55f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872651 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/66d09685-dfb3-4344-8cb2-f4e820e05bac-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6rh7z\" (UID: \"66d09685-dfb3-4344-8cb2-f4e820e05bac\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872674 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x5dn\" (UniqueName: \"kubernetes.io/projected/30e3e22f-c17f-4efa-b5cd-33406a09def1-kube-api-access-7x5dn\") pod \"service-ca-9c57cc56f-d948p\" (UID: \"30e3e22f-c17f-4efa-b5cd-33406a09def1\") " pod="openshift-service-ca/service-ca-9c57cc56f-d948p" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872695 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/821984e5-b740-4332-9c06-1ccefda9a9fb-serving-cert\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872711 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/821984e5-b740-4332-9c06-1ccefda9a9fb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872744 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7bz7\" (UniqueName: \"kubernetes.io/projected/86a63698-caa8-4173-8f86-1bb56087babb-kube-api-access-f7bz7\") pod \"openshift-controller-manager-operator-756b6f6bc6-49jn7\" (UID: \"86a63698-caa8-4173-8f86-1bb56087babb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-49jn7" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872763 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c2c0ca2-4182-4c24-8710-c3b632e768f9-serving-cert\") pod \"console-operator-58897d9998-f95s6\" (UID: \"7c2c0ca2-4182-4c24-8710-c3b632e768f9\") " pod="openshift-console-operator/console-operator-58897d9998-f95s6" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872781 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffpfg\" (UniqueName: \"kubernetes.io/projected/3cb9bb68-546e-4e52-9a15-5aeee129a5b4-kube-api-access-ffpfg\") pod \"downloads-7954f5f757-bzps5\" (UID: \"3cb9bb68-546e-4e52-9a15-5aeee129a5b4\") " pod="openshift-console/downloads-7954f5f757-bzps5" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872810 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872834 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7f9f078-ff94-4adc-be45-927915d6e55f-serving-cert\") pod \"etcd-operator-b45778765-d9d4k\" (UID: \"a7f9f078-ff94-4adc-be45-927915d6e55f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872865 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/66d09685-dfb3-4344-8cb2-f4e820e05bac-proxy-tls\") pod \"machine-config-operator-74547568cd-6rh7z\" (UID: \"66d09685-dfb3-4344-8cb2-f4e820e05bac\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872898 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66jvb\" (UniqueName: \"kubernetes.io/projected/de09ed81-3849-460e-93f2-fb0f5d72a4b9-kube-api-access-66jvb\") pod \"packageserver-d55dfcdfc-5vpqp\" (UID: \"de09ed81-3849-460e-93f2-fb0f5d72a4b9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872915 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/fdb5ad89-a263-4f1f-bbe5-42bd5b91ea25-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-g25mq\" (UID: \"fdb5ad89-a263-4f1f-bbe5-42bd5b91ea25\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g25mq" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872929 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2201ed6d-9862-43d5-98ed-949539d6d0ae-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ql7rx\" (UID: \"2201ed6d-9862-43d5-98ed-949539d6d0ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ql7rx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872966 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.872984 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3253b15-b01e-4c6b-91df-578ddb39b234-serving-cert\") pod \"authentication-operator-69f744f599-jq824\" (UID: \"e3253b15-b01e-4c6b-91df-578ddb39b234\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.873002 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q99f\" (UniqueName: \"kubernetes.io/projected/7c2c0ca2-4182-4c24-8710-c3b632e768f9-kube-api-access-7q99f\") pod \"console-operator-58897d9998-f95s6\" (UID: \"7c2c0ca2-4182-4c24-8710-c3b632e768f9\") " pod="openshift-console-operator/console-operator-58897d9998-f95s6" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.873045 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7c2c0ca2-4182-4c24-8710-c3b632e768f9-trusted-ca\") pod \"console-operator-58897d9998-f95s6\" (UID: \"7c2c0ca2-4182-4c24-8710-c3b632e768f9\") " pod="openshift-console-operator/console-operator-58897d9998-f95s6" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.873084 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6q44\" (UniqueName: \"kubernetes.io/projected/821984e5-b740-4332-9c06-1ccefda9a9fb-kube-api-access-v6q44\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.873103 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ddd4a23a-8856-480c-91d6-829f766120a9-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-vdl4w\" (UID: \"ddd4a23a-8856-480c-91d6-829f766120a9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.873136 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5k6l\" (UniqueName: \"kubernetes.io/projected/967509da-8c27-49be-aa96-3a7129937ddd-kube-api-access-m5k6l\") pod \"openshift-apiserver-operator-796bbdcf4f-7t272\" (UID: \"967509da-8c27-49be-aa96-3a7129937ddd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7t272" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.873230 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.873286 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3253b15-b01e-4c6b-91df-578ddb39b234-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jq824\" (UID: \"e3253b15-b01e-4c6b-91df-578ddb39b234\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.874518 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/345bb34c-a9b2-4e88-94ce-3187b8933057-serving-cert\") pod \"openshift-config-operator-7777fb866f-4kpjx\" (UID: \"345bb34c-a9b2-4e88-94ce-3187b8933057\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.874546 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/821984e5-b740-4332-9c06-1ccefda9a9fb-audit-policies\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.874592 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/821984e5-b740-4332-9c06-1ccefda9a9fb-encryption-config\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.874648 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b42c5242-55b6-43aa-a9bf-e368d446c3a5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qsppc\" (UID: \"b42c5242-55b6-43aa-a9bf-e368d446c3a5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qsppc" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.874668 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/de09ed81-3849-460e-93f2-fb0f5d72a4b9-tmpfs\") pod \"packageserver-d55dfcdfc-5vpqp\" (UID: \"de09ed81-3849-460e-93f2-fb0f5d72a4b9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.874880 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e831e54b-3826-4415-9b43-ce02e5ed81bf-audit-dir\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.874920 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b42c5242-55b6-43aa-a9bf-e368d446c3a5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qsppc\" (UID: \"b42c5242-55b6-43aa-a9bf-e368d446c3a5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qsppc" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.874944 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cszmb\" (UniqueName: \"kubernetes.io/projected/2201ed6d-9862-43d5-98ed-949539d6d0ae-kube-api-access-cszmb\") pod \"kube-storage-version-migrator-operator-b67b599dd-ql7rx\" (UID: \"2201ed6d-9862-43d5-98ed-949539d6d0ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ql7rx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.874974 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-audit-policies\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.875017 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.875040 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a7f9f078-ff94-4adc-be45-927915d6e55f-etcd-service-ca\") pod \"etcd-operator-b45778765-d9d4k\" (UID: \"a7f9f078-ff94-4adc-be45-927915d6e55f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.875058 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3253b15-b01e-4c6b-91df-578ddb39b234-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jq824\" (UID: \"e3253b15-b01e-4c6b-91df-578ddb39b234\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.875153 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ddd4a23a-8856-480c-91d6-829f766120a9-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-vdl4w\" (UID: \"ddd4a23a-8856-480c-91d6-829f766120a9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.875175 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/66d09685-dfb3-4344-8cb2-f4e820e05bac-images\") pod \"machine-config-operator-74547568cd-6rh7z\" (UID: \"66d09685-dfb3-4344-8cb2-f4e820e05bac\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.875212 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/821984e5-b740-4332-9c06-1ccefda9a9fb-etcd-client\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.875246 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mpwg\" (UniqueName: \"kubernetes.io/projected/e3253b15-b01e-4c6b-91df-578ddb39b234-kube-api-access-8mpwg\") pod \"authentication-operator-69f744f599-jq824\" (UID: \"e3253b15-b01e-4c6b-91df-578ddb39b234\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.875293 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/de09ed81-3849-460e-93f2-fb0f5d72a4b9-apiservice-cert\") pod \"packageserver-d55dfcdfc-5vpqp\" (UID: \"de09ed81-3849-460e-93f2-fb0f5d72a4b9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.875313 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/b9b87214-6f46-4d58-b2fe-5618cfe4e0f0-machine-approver-tls\") pod \"machine-approver-56656f9798-pcc6z\" (UID: \"b9b87214-6f46-4d58-b2fe-5618cfe4e0f0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pcc6z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.875789 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a7f9f078-ff94-4adc-be45-927915d6e55f-etcd-ca\") pod \"etcd-operator-b45778765-d9d4k\" (UID: \"a7f9f078-ff94-4adc-be45-927915d6e55f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.875828 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.875876 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a7f9f078-ff94-4adc-be45-927915d6e55f-etcd-client\") pod \"etcd-operator-b45778765-d9d4k\" (UID: \"a7f9f078-ff94-4adc-be45-927915d6e55f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.875901 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzmbq\" (UniqueName: \"kubernetes.io/projected/fdb5ad89-a263-4f1f-bbe5-42bd5b91ea25-kube-api-access-vzmbq\") pod \"cluster-samples-operator-665b6dd947-g25mq\" (UID: \"fdb5ad89-a263-4f1f-bbe5-42bd5b91ea25\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g25mq" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.875930 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b89659e-58f4-4c11-9839-2be0bdcd6e5e-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5pmk4\" (UID: \"8b89659e-58f4-4c11-9839-2be0bdcd6e5e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.875951 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.875971 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.875998 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c2c0ca2-4182-4c24-8710-c3b632e768f9-config\") pod \"console-operator-58897d9998-f95s6\" (UID: \"7c2c0ca2-4182-4c24-8710-c3b632e768f9\") " pod="openshift-console-operator/console-operator-58897d9998-f95s6" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.876019 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/821984e5-b740-4332-9c06-1ccefda9a9fb-audit-dir\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.876041 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.876059 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq4bg\" (UniqueName: \"kubernetes.io/projected/8b89659e-58f4-4c11-9839-2be0bdcd6e5e-kube-api-access-nq4bg\") pod \"ingress-operator-5b745b69d9-5pmk4\" (UID: \"8b89659e-58f4-4c11-9839-2be0bdcd6e5e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.876087 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpfx4\" (UniqueName: \"kubernetes.io/projected/5a537ec4-ab86-4bc9-81d2-3870472b767a-kube-api-access-dpfx4\") pod \"controller-manager-879f6c89f-5lwz5\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.876106 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/821984e5-b740-4332-9c06-1ccefda9a9fb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.876128 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/de09ed81-3849-460e-93f2-fb0f5d72a4b9-webhook-cert\") pod \"packageserver-d55dfcdfc-5vpqp\" (UID: \"de09ed81-3849-460e-93f2-fb0f5d72a4b9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.876152 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/345bb34c-a9b2-4e88-94ce-3187b8933057-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4kpjx\" (UID: \"345bb34c-a9b2-4e88-94ce-3187b8933057\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.876173 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/30e3e22f-c17f-4efa-b5cd-33406a09def1-signing-key\") pod \"service-ca-9c57cc56f-d948p\" (UID: \"30e3e22f-c17f-4efa-b5cd-33406a09def1\") " pod="openshift-service-ca/service-ca-9c57cc56f-d948p" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.876205 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/30e3e22f-c17f-4efa-b5cd-33406a09def1-signing-cabundle\") pod \"service-ca-9c57cc56f-d948p\" (UID: \"30e3e22f-c17f-4efa-b5cd-33406a09def1\") " pod="openshift-service-ca/service-ca-9c57cc56f-d948p" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.876224 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/967509da-8c27-49be-aa96-3a7129937ddd-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7t272\" (UID: \"967509da-8c27-49be-aa96-3a7129937ddd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7t272" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.876243 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b89659e-58f4-4c11-9839-2be0bdcd6e5e-trusted-ca\") pod \"ingress-operator-5b745b69d9-5pmk4\" (UID: \"8b89659e-58f4-4c11-9839-2be0bdcd6e5e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.876262 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ddd4a23a-8856-480c-91d6-829f766120a9-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-vdl4w\" (UID: \"ddd4a23a-8856-480c-91d6-829f766120a9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.876281 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/967509da-8c27-49be-aa96-3a7129937ddd-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7t272\" (UID: \"967509da-8c27-49be-aa96-3a7129937ddd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7t272" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.876304 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdv8b\" (UniqueName: \"kubernetes.io/projected/b9b87214-6f46-4d58-b2fe-5618cfe4e0f0-kube-api-access-jdv8b\") pod \"machine-approver-56656f9798-pcc6z\" (UID: \"b9b87214-6f46-4d58-b2fe-5618cfe4e0f0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pcc6z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.876322 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsk2p\" (UniqueName: \"kubernetes.io/projected/ddd4a23a-8856-480c-91d6-829f766120a9-kube-api-access-wsk2p\") pod \"cluster-image-registry-operator-dc59b4c8b-vdl4w\" (UID: \"ddd4a23a-8856-480c-91d6-829f766120a9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.876348 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhb2k\" (UniqueName: \"kubernetes.io/projected/e831e54b-3826-4415-9b43-ce02e5ed81bf-kube-api-access-nhb2k\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.876372 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/11f24e26-e608-47a1-b8b1-8ac0159c1c5a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-fqtd8\" (UID: \"11f24e26-e608-47a1-b8b1-8ac0159c1c5a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fqtd8" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.876391 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.879541 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8b89659e-58f4-4c11-9839-2be0bdcd6e5e-metrics-tls\") pod \"ingress-operator-5b745b69d9-5pmk4\" (UID: \"8b89659e-58f4-4c11-9839-2be0bdcd6e5e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.879624 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.882619 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ql7rx"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.884208 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a7f9f078-ff94-4adc-be45-927915d6e55f-etcd-ca\") pod \"etcd-operator-b45778765-d9d4k\" (UID: \"a7f9f078-ff94-4adc-be45-927915d6e55f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.887743 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a7f9f078-ff94-4adc-be45-927915d6e55f-etcd-client\") pod \"etcd-operator-b45778765-d9d4k\" (UID: \"a7f9f078-ff94-4adc-be45-927915d6e55f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.888617 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c2c0ca2-4182-4c24-8710-c3b632e768f9-config\") pod \"console-operator-58897d9998-f95s6\" (UID: \"7c2c0ca2-4182-4c24-8710-c3b632e768f9\") " pod="openshift-console-operator/console-operator-58897d9998-f95s6" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.888670 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/821984e5-b740-4332-9c06-1ccefda9a9fb-audit-dir\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.889318 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/821984e5-b740-4332-9c06-1ccefda9a9fb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.889758 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/821984e5-b740-4332-9c06-1ccefda9a9fb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.890609 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/fdb5ad89-a263-4f1f-bbe5-42bd5b91ea25-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-g25mq\" (UID: \"fdb5ad89-a263-4f1f-bbe5-42bd5b91ea25\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g25mq" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.895058 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/de09ed81-3849-460e-93f2-fb0f5d72a4b9-tmpfs\") pod \"packageserver-d55dfcdfc-5vpqp\" (UID: \"de09ed81-3849-460e-93f2-fb0f5d72a4b9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.895749 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/821984e5-b740-4332-9c06-1ccefda9a9fb-audit-policies\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.895763 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3253b15-b01e-4c6b-91df-578ddb39b234-service-ca-bundle\") pod \"authentication-operator-69f744f599-jq824\" (UID: \"e3253b15-b01e-4c6b-91df-578ddb39b234\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.895858 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2201ed6d-9862-43d5-98ed-949539d6d0ae-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ql7rx\" (UID: \"2201ed6d-9862-43d5-98ed-949539d6d0ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ql7rx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.896265 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7c2c0ca2-4182-4c24-8710-c3b632e768f9-trusted-ca\") pod \"console-operator-58897d9998-f95s6\" (UID: \"7c2c0ca2-4182-4c24-8710-c3b632e768f9\") " pod="openshift-console-operator/console-operator-58897d9998-f95s6" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.896681 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-qjrkr"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.896393 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2201ed6d-9862-43d5-98ed-949539d6d0ae-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ql7rx\" (UID: \"2201ed6d-9862-43d5-98ed-949539d6d0ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ql7rx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.899390 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/345bb34c-a9b2-4e88-94ce-3187b8933057-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4kpjx\" (UID: \"345bb34c-a9b2-4e88-94ce-3187b8933057\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.900085 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c2c0ca2-4182-4c24-8710-c3b632e768f9-serving-cert\") pod \"console-operator-58897d9998-f95s6\" (UID: \"7c2c0ca2-4182-4c24-8710-c3b632e768f9\") " pod="openshift-console-operator/console-operator-58897d9998-f95s6" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.900470 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/345bb34c-a9b2-4e88-94ce-3187b8933057-serving-cert\") pod \"openshift-config-operator-7777fb866f-4kpjx\" (UID: \"345bb34c-a9b2-4e88-94ce-3187b8933057\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.904132 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.904248 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/66d09685-dfb3-4344-8cb2-f4e820e05bac-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6rh7z\" (UID: \"66d09685-dfb3-4344-8cb2-f4e820e05bac\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.904307 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/de09ed81-3849-460e-93f2-fb0f5d72a4b9-webhook-cert\") pod \"packageserver-d55dfcdfc-5vpqp\" (UID: \"de09ed81-3849-460e-93f2-fb0f5d72a4b9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.904750 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/30e3e22f-c17f-4efa-b5cd-33406a09def1-signing-cabundle\") pod \"service-ca-9c57cc56f-d948p\" (UID: \"30e3e22f-c17f-4efa-b5cd-33406a09def1\") " pod="openshift-service-ca/service-ca-9c57cc56f-d948p" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.905789 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7f9f078-ff94-4adc-be45-927915d6e55f-config\") pod \"etcd-operator-b45778765-d9d4k\" (UID: \"a7f9f078-ff94-4adc-be45-927915d6e55f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.906124 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.906265 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3253b15-b01e-4c6b-91df-578ddb39b234-serving-cert\") pod \"authentication-operator-69f744f599-jq824\" (UID: \"e3253b15-b01e-4c6b-91df-578ddb39b234\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.907282 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3253b15-b01e-4c6b-91df-578ddb39b234-config\") pod \"authentication-operator-69f744f599-jq824\" (UID: \"e3253b15-b01e-4c6b-91df-578ddb39b234\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.907438 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5a537ec4-ab86-4bc9-81d2-3870472b767a-client-ca\") pod \"controller-manager-879f6c89f-5lwz5\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.907784 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.908065 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a537ec4-ab86-4bc9-81d2-3870472b767a-config\") pod \"controller-manager-879f6c89f-5lwz5\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.908778 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b9b87214-6f46-4d58-b2fe-5618cfe4e0f0-auth-proxy-config\") pod \"machine-approver-56656f9798-pcc6z\" (UID: \"b9b87214-6f46-4d58-b2fe-5618cfe4e0f0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pcc6z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.909289 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5a537ec4-ab86-4bc9-81d2-3870472b767a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-5lwz5\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.909493 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b42c5242-55b6-43aa-a9bf-e368d446c3a5-config\") pod \"kube-controller-manager-operator-78b949d7b-qsppc\" (UID: \"b42c5242-55b6-43aa-a9bf-e368d446c3a5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qsppc" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.910037 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ddd4a23a-8856-480c-91d6-829f766120a9-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-vdl4w\" (UID: \"ddd4a23a-8856-480c-91d6-829f766120a9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.910172 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/66d09685-dfb3-4344-8cb2-f4e820e05bac-images\") pod \"machine-config-operator-74547568cd-6rh7z\" (UID: \"66d09685-dfb3-4344-8cb2-f4e820e05bac\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.910451 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b89659e-58f4-4c11-9839-2be0bdcd6e5e-trusted-ca\") pod \"ingress-operator-5b745b69d9-5pmk4\" (UID: \"8b89659e-58f4-4c11-9839-2be0bdcd6e5e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.912775 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9b87214-6f46-4d58-b2fe-5618cfe4e0f0-config\") pod \"machine-approver-56656f9798-pcc6z\" (UID: \"b9b87214-6f46-4d58-b2fe-5618cfe4e0f0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pcc6z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.912803 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7f9f078-ff94-4adc-be45-927915d6e55f-serving-cert\") pod \"etcd-operator-b45778765-d9d4k\" (UID: \"a7f9f078-ff94-4adc-be45-927915d6e55f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.913171 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/967509da-8c27-49be-aa96-3a7129937ddd-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7t272\" (UID: \"967509da-8c27-49be-aa96-3a7129937ddd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7t272" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.913461 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a7f9f078-ff94-4adc-be45-927915d6e55f-etcd-service-ca\") pod \"etcd-operator-b45778765-d9d4k\" (UID: \"a7f9f078-ff94-4adc-be45-927915d6e55f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.913600 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/821984e5-b740-4332-9c06-1ccefda9a9fb-encryption-config\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.914156 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a537ec4-ab86-4bc9-81d2-3870472b767a-serving-cert\") pod \"controller-manager-879f6c89f-5lwz5\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.914795 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/821984e5-b740-4332-9c06-1ccefda9a9fb-etcd-client\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.914863 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-2q49s"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.915285 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ddd4a23a-8856-480c-91d6-829f766120a9-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-vdl4w\" (UID: \"ddd4a23a-8856-480c-91d6-829f766120a9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.915432 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2q49s" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.916088 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-6wnjb"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.916497 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-6wnjb" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.917201 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/967509da-8c27-49be-aa96-3a7129937ddd-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7t272\" (UID: \"967509da-8c27-49be-aa96-3a7129937ddd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7t272" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.917239 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/30e3e22f-c17f-4efa-b5cd-33406a09def1-signing-key\") pod \"service-ca-9c57cc56f-d948p\" (UID: \"30e3e22f-c17f-4efa-b5cd-33406a09def1\") " pod="openshift-service-ca/service-ca-9c57cc56f-d948p" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.917547 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/821984e5-b740-4332-9c06-1ccefda9a9fb-serving-cert\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.917933 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/de09ed81-3849-460e-93f2-fb0f5d72a4b9-apiservice-cert\") pod \"packageserver-d55dfcdfc-5vpqp\" (UID: \"de09ed81-3849-460e-93f2-fb0f5d72a4b9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.918186 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/66d09685-dfb3-4344-8cb2-f4e820e05bac-proxy-tls\") pod \"machine-config-operator-74547568cd-6rh7z\" (UID: \"66d09685-dfb3-4344-8cb2-f4e820e05bac\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.918432 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-qhldd"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.919220 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.919626 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b42c5242-55b6-43aa-a9bf-e368d446c3a5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qsppc\" (UID: \"b42c5242-55b6-43aa-a9bf-e368d446c3a5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qsppc" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.921362 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/b9b87214-6f46-4d58-b2fe-5618cfe4e0f0-machine-approver-tls\") pod \"machine-approver-56656f9798-pcc6z\" (UID: \"b9b87214-6f46-4d58-b2fe-5618cfe4e0f0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pcc6z" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.922597 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g25mq"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.925778 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.926236 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4n756"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.928387 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.929616 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjcqb"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.930128 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/11f24e26-e608-47a1-b8b1-8ac0159c1c5a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-fqtd8\" (UID: \"11f24e26-e608-47a1-b8b1-8ac0159c1c5a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fqtd8" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.930755 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-bzps5"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.931783 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-t8ggt"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.932824 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.933956 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-x4nlj"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.935033 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qsppc"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.937980 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.940090 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.941983 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-49jn7"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.946028 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.946392 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-2gsdv"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.949717 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fqtd8"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.950550 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.952011 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9pzkg"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.953266 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-f95s6"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.955836 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c5fmb"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.957106 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-fjbmw"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.957813 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-fjbmw" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.959182 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bfn42"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.960045 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-7c6cp"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.961289 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zbh4h"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.962684 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-fjbmw"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.964736 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-qhldd"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.965604 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2q49s"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.966070 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.966804 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hfw59"] Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.976918 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86a63698-caa8-4173-8f86-1bb56087babb-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-49jn7\" (UID: \"86a63698-caa8-4173-8f86-1bb56087babb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-49jn7" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.976948 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86a63698-caa8-4173-8f86-1bb56087babb-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-49jn7\" (UID: \"86a63698-caa8-4173-8f86-1bb56087babb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-49jn7" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.976969 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.977054 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7bz7\" (UniqueName: \"kubernetes.io/projected/86a63698-caa8-4173-8f86-1bb56087babb-kube-api-access-f7bz7\") pod \"openshift-controller-manager-operator-756b6f6bc6-49jn7\" (UID: \"86a63698-caa8-4173-8f86-1bb56087babb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-49jn7" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.977088 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.977203 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.977275 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.977332 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e831e54b-3826-4415-9b43-ce02e5ed81bf-audit-dir\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.977373 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-audit-policies\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.977412 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.977465 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.977508 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.977536 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.977561 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.977640 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.977667 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhb2k\" (UniqueName: \"kubernetes.io/projected/e831e54b-3826-4415-9b43-ce02e5ed81bf-kube-api-access-nhb2k\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.977692 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.978239 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e831e54b-3826-4415-9b43-ce02e5ed81bf-audit-dir\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:27 crc kubenswrapper[4904]: I0214 11:12:27.985910 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.006018 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.025352 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.046735 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.065993 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.086367 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.106043 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.125729 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.145366 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.166424 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.186715 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.206008 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.226371 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.231380 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86a63698-caa8-4173-8f86-1bb56087babb-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-49jn7\" (UID: \"86a63698-caa8-4173-8f86-1bb56087babb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-49jn7" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.245673 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.266278 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.269133 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86a63698-caa8-4173-8f86-1bb56087babb-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-49jn7\" (UID: \"86a63698-caa8-4173-8f86-1bb56087babb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-49jn7" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.286226 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.305385 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.326278 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.345512 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.366265 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.385355 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.405868 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.425458 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.445470 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.466629 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.486128 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.506608 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.525657 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.545930 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.566624 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.585189 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.605937 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.626072 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.645793 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.668120 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.686509 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.706005 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.725999 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.746291 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.763961 4904 request.go:700] Waited for 1.009922231s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/secrets?fieldSelector=metadata.name%3Doauth-openshift-dockercfg-znhcc&limit=500&resourceVersion=0 Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.767183 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.786208 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.792242 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.821160 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.825623 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.831740 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.832328 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.845960 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.864368 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.867130 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.875327 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.886489 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.897178 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.911320 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.921962 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.926291 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.930890 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.946120 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.968258 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.968903 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-audit-policies\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:28 crc kubenswrapper[4904]: E0214 11:12:28.978574 4904 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Feb 14 11:12:28 crc kubenswrapper[4904]: E0214 11:12:28.978679 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-trusted-ca-bundle podName:e831e54b-3826-4415-9b43-ce02e5ed81bf nodeName:}" failed. No retries permitted until 2026-02-14 11:12:29.478658335 +0000 UTC m=+140.291422996 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-trusted-ca-bundle") pod "oauth-openshift-558db77b4-bfn42" (UID: "e831e54b-3826-4415-9b43-ce02e5ed81bf") : failed to sync configmap cache: timed out waiting for the condition Feb 14 11:12:28 crc kubenswrapper[4904]: E0214 11:12:28.978582 4904 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-service-ca: failed to sync configmap cache: timed out waiting for the condition Feb 14 11:12:28 crc kubenswrapper[4904]: E0214 11:12:28.979012 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-service-ca podName:e831e54b-3826-4415-9b43-ce02e5ed81bf nodeName:}" failed. No retries permitted until 2026-02-14 11:12:29.479003133 +0000 UTC m=+140.291767794 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-service-ca" (UniqueName: "kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-service-ca") pod "oauth-openshift-558db77b4-bfn42" (UID: "e831e54b-3826-4415-9b43-ce02e5ed81bf") : failed to sync configmap cache: timed out waiting for the condition Feb 14 11:12:28 crc kubenswrapper[4904]: E0214 11:12:28.978589 4904 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-cliconfig: failed to sync configmap cache: timed out waiting for the condition Feb 14 11:12:28 crc kubenswrapper[4904]: E0214 11:12:28.979043 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-cliconfig podName:e831e54b-3826-4415-9b43-ce02e5ed81bf nodeName:}" failed. No retries permitted until 2026-02-14 11:12:29.479037184 +0000 UTC m=+140.291801845 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-cliconfig" (UniqueName: "kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-cliconfig") pod "oauth-openshift-558db77b4-bfn42" (UID: "e831e54b-3826-4415-9b43-ce02e5ed81bf") : failed to sync configmap cache: timed out waiting for the condition Feb 14 11:12:28 crc kubenswrapper[4904]: I0214 11:12:28.986340 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.006221 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.031569 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.045604 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.065354 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.085902 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.106057 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.126150 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.145971 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.165455 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.186326 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.206117 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.225604 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.245443 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.265879 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.293684 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.305108 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.325301 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.351780 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.366123 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.386756 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.406508 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.444715 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n56s2\" (UniqueName: \"kubernetes.io/projected/800acf17-404a-4dc3-a25f-66e779d292fd-kube-api-access-n56s2\") pod \"machine-api-operator-5694c8668f-hgggh\" (UID: \"800acf17-404a-4dc3-a25f-66e779d292fd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hgggh" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.460628 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfr2m\" (UniqueName: \"kubernetes.io/projected/80b9d666-fa50-4d49-8c3f-efbd9d1247a9-kube-api-access-qfr2m\") pod \"apiserver-76f77b778f-t8ggt\" (UID: \"80b9d666-fa50-4d49-8c3f-efbd9d1247a9\") " pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.486093 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.495954 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.496156 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.497379 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.497304 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.496877 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.498380 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.513368 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-hgggh" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.536024 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x5dn\" (UniqueName: \"kubernetes.io/projected/30e3e22f-c17f-4efa-b5cd-33406a09def1-kube-api-access-7x5dn\") pod \"service-ca-9c57cc56f-d948p\" (UID: \"30e3e22f-c17f-4efa-b5cd-33406a09def1\") " pod="openshift-service-ca/service-ca-9c57cc56f-d948p" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.550297 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzmbq\" (UniqueName: \"kubernetes.io/projected/fdb5ad89-a263-4f1f-bbe5-42bd5b91ea25-kube-api-access-vzmbq\") pod \"cluster-samples-operator-665b6dd947-g25mq\" (UID: \"fdb5ad89-a263-4f1f-bbe5-42bd5b91ea25\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g25mq" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.560649 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b89659e-58f4-4c11-9839-2be0bdcd6e5e-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5pmk4\" (UID: \"8b89659e-58f4-4c11-9839-2be0bdcd6e5e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.580848 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq4bg\" (UniqueName: \"kubernetes.io/projected/8b89659e-58f4-4c11-9839-2be0bdcd6e5e-kube-api-access-nq4bg\") pod \"ingress-operator-5b745b69d9-5pmk4\" (UID: \"8b89659e-58f4-4c11-9839-2be0bdcd6e5e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.606888 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpfx4\" (UniqueName: \"kubernetes.io/projected/5a537ec4-ab86-4bc9-81d2-3870472b767a-kube-api-access-dpfx4\") pod \"controller-manager-879f6c89f-5lwz5\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.614514 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.640246 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q99f\" (UniqueName: \"kubernetes.io/projected/7c2c0ca2-4182-4c24-8710-c3b632e768f9-kube-api-access-7q99f\") pod \"console-operator-58897d9998-f95s6\" (UID: \"7c2c0ca2-4182-4c24-8710-c3b632e768f9\") " pod="openshift-console-operator/console-operator-58897d9998-f95s6" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.644303 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g25mq" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.657315 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66jvb\" (UniqueName: \"kubernetes.io/projected/de09ed81-3849-460e-93f2-fb0f5d72a4b9-kube-api-access-66jvb\") pod \"packageserver-d55dfcdfc-5vpqp\" (UID: \"de09ed81-3849-460e-93f2-fb0f5d72a4b9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.660744 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffpfg\" (UniqueName: \"kubernetes.io/projected/3cb9bb68-546e-4e52-9a15-5aeee129a5b4-kube-api-access-ffpfg\") pod \"downloads-7954f5f757-bzps5\" (UID: \"3cb9bb68-546e-4e52-9a15-5aeee129a5b4\") " pod="openshift-console/downloads-7954f5f757-bzps5" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.661399 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.680768 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b42c5242-55b6-43aa-a9bf-e368d446c3a5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qsppc\" (UID: \"b42c5242-55b6-43aa-a9bf-e368d446c3a5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qsppc" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.702604 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6q44\" (UniqueName: \"kubernetes.io/projected/821984e5-b740-4332-9c06-1ccefda9a9fb-kube-api-access-v6q44\") pod \"apiserver-7bbb656c7d-jrpqw\" (UID: \"821984e5-b740-4332-9c06-1ccefda9a9fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.707736 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qsppc" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.719200 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5k6l\" (UniqueName: \"kubernetes.io/projected/967509da-8c27-49be-aa96-3a7129937ddd-kube-api-access-m5k6l\") pod \"openshift-apiserver-operator-796bbdcf4f-7t272\" (UID: \"967509da-8c27-49be-aa96-3a7129937ddd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7t272" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.736739 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.747468 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cszmb\" (UniqueName: \"kubernetes.io/projected/2201ed6d-9862-43d5-98ed-949539d6d0ae-kube-api-access-cszmb\") pod \"kube-storage-version-migrator-operator-b67b599dd-ql7rx\" (UID: \"2201ed6d-9862-43d5-98ed-949539d6d0ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ql7rx" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.762549 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vjcd\" (UniqueName: \"kubernetes.io/projected/a7f9f078-ff94-4adc-be45-927915d6e55f-kube-api-access-6vjcd\") pod \"etcd-operator-b45778765-d9d4k\" (UID: \"a7f9f078-ff94-4adc-be45-927915d6e55f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.762697 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.768534 4904 request.go:700] Waited for 1.864794597s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/serviceaccounts/olm-operator-serviceaccount/token Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.769984 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-hgggh"] Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.782806 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.783266 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-d948p" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.787426 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ql7rx" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.799742 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcq6t\" (UniqueName: \"kubernetes.io/projected/66d09685-dfb3-4344-8cb2-f4e820e05bac-kube-api-access-pcq6t\") pod \"machine-config-operator-74547568cd-6rh7z\" (UID: \"66d09685-dfb3-4344-8cb2-f4e820e05bac\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.821878 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7vch\" (UniqueName: \"kubernetes.io/projected/345bb34c-a9b2-4e88-94ce-3187b8933057-kube-api-access-c7vch\") pod \"openshift-config-operator-7777fb866f-4kpjx\" (UID: \"345bb34c-a9b2-4e88-94ce-3187b8933057\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.822906 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sp79\" (UniqueName: \"kubernetes.io/projected/11f24e26-e608-47a1-b8b1-8ac0159c1c5a-kube-api-access-2sp79\") pod \"package-server-manager-789f6589d5-fqtd8\" (UID: \"11f24e26-e608-47a1-b8b1-8ac0159c1c5a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fqtd8" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.843780 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.856452 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mpwg\" (UniqueName: \"kubernetes.io/projected/e3253b15-b01e-4c6b-91df-578ddb39b234-kube-api-access-8mpwg\") pod \"authentication-operator-69f744f599-jq824\" (UID: \"e3253b15-b01e-4c6b-91df-578ddb39b234\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.860679 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.888594 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ddd4a23a-8856-480c-91d6-829f766120a9-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-vdl4w\" (UID: \"ddd4a23a-8856-480c-91d6-829f766120a9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.904212 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-f95s6" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.907228 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.907560 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsk2p\" (UniqueName: \"kubernetes.io/projected/ddd4a23a-8856-480c-91d6-829f766120a9-kube-api-access-wsk2p\") pod \"cluster-image-registry-operator-dc59b4c8b-vdl4w\" (UID: \"ddd4a23a-8856-480c-91d6-829f766120a9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.914493 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-t8ggt"] Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.926202 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-bzps5" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.929116 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.945569 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5lwz5"] Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.948230 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.948449 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdv8b\" (UniqueName: \"kubernetes.io/projected/b9b87214-6f46-4d58-b2fe-5618cfe4e0f0-kube-api-access-jdv8b\") pod \"machine-approver-56656f9798-pcc6z\" (UID: \"b9b87214-6f46-4d58-b2fe-5618cfe4e0f0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pcc6z" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.965602 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 14 11:12:29 crc kubenswrapper[4904]: I0214 11:12:29.987198 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.006398 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.015120 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7t272" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.024732 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.027148 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.047235 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.048968 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.054497 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pcc6z" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.066413 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.068891 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fqtd8" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.088736 4904 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.111328 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.128910 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.147204 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.186024 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7bz7\" (UniqueName: \"kubernetes.io/projected/86a63698-caa8-4173-8f86-1bb56087babb-kube-api-access-f7bz7\") pod \"openshift-controller-manager-operator-756b6f6bc6-49jn7\" (UID: \"86a63698-caa8-4173-8f86-1bb56087babb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-49jn7" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.202688 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4"] Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.207506 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhb2k\" (UniqueName: \"kubernetes.io/projected/e831e54b-3826-4415-9b43-ce02e5ed81bf-kube-api-access-nhb2k\") pod \"oauth-openshift-558db77b4-bfn42\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:30 crc kubenswrapper[4904]: W0214 11:12:30.212706 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9b87214_6f46_4d58_b2fe_5618cfe4e0f0.slice/crio-27a6017f2fc321b21d2b3c7101b553d8af1a06a2908e1fdee558e916c1651522 WatchSource:0}: Error finding container 27a6017f2fc321b21d2b3c7101b553d8af1a06a2908e1fdee558e916c1651522: Status 404 returned error can't find the container with id 27a6017f2fc321b21d2b3c7101b553d8af1a06a2908e1fdee558e916c1651522 Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.235895 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g25mq"] Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.311613 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qsppc"] Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.318045 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/477826a2-46cd-4145-afc9-92b69930526e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-c5fmb\" (UID: \"477826a2-46cd-4145-afc9-92b69930526e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c5fmb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.318172 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d97a9335-cf15-4b30-ba2e-846e0daed72d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.318291 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/261ad318-57b3-48ca-8dac-e8bd53528694-console-serving-cert\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320108 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d77fde-3517-4aea-abcb-055541763e48-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4n756\" (UID: \"e9d77fde-3517-4aea-abcb-055541763e48\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4n756" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320141 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/477826a2-46cd-4145-afc9-92b69930526e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-c5fmb\" (UID: \"477826a2-46cd-4145-afc9-92b69930526e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c5fmb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320167 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d-proxy-tls\") pod \"machine-config-controller-84d6567774-qjrkr\" (UID: \"fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qjrkr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320188 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbd4ae38-0244-445f-a4e8-6bffc1714d2d-config\") pod \"service-ca-operator-777779d784-hfw59\" (UID: \"bbd4ae38-0244-445f-a4e8-6bffc1714d2d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hfw59" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320234 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/652a99d8-a666-4c63-8bc5-71953e836b01-srv-cert\") pod \"catalog-operator-68c6474976-tbxbl\" (UID: \"652a99d8-a666-4c63-8bc5-71953e836b01\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320287 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f130a8a5-24b3-4916-a0c7-b1fa5e55cad4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9pzkg\" (UID: \"f130a8a5-24b3-4916-a0c7-b1fa5e55cad4\") " pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320309 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfn4h\" (UniqueName: \"kubernetes.io/projected/652a99d8-a666-4c63-8bc5-71953e836b01-kube-api-access-pfn4h\") pod \"catalog-operator-68c6474976-tbxbl\" (UID: \"652a99d8-a666-4c63-8bc5-71953e836b01\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320335 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/91b7cb2e-1781-41da-9c90-13b40be0a2b4-stats-auth\") pod \"router-default-5444994796-qh6sr\" (UID: \"91b7cb2e-1781-41da-9c90-13b40be0a2b4\") " pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320356 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b1710cb1-bd05-4970-9b17-95cf75b9aec4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-x4nlj\" (UID: \"b1710cb1-bd05-4970-9b17-95cf75b9aec4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-x4nlj" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320431 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d97a9335-cf15-4b30-ba2e-846e0daed72d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320467 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d910fb92-a40a-4167-be6f-113f22c206b2-client-ca\") pod \"route-controller-manager-6576b87f9c-rnf52\" (UID: \"d910fb92-a40a-4167-be6f-113f22c206b2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320489 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzzg8\" (UniqueName: \"kubernetes.io/projected/f130a8a5-24b3-4916-a0c7-b1fa5e55cad4-kube-api-access-pzzg8\") pod \"marketplace-operator-79b997595-9pzkg\" (UID: \"f130a8a5-24b3-4916-a0c7-b1fa5e55cad4\") " pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320512 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ca75271d-7ae7-4a71-af10-732866e33d74-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sjcqb\" (UID: \"ca75271d-7ae7-4a71-af10-732866e33d74\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjcqb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320580 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d77fde-3517-4aea-abcb-055541763e48-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4n756\" (UID: \"e9d77fde-3517-4aea-abcb-055541763e48\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4n756" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320602 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d97a9335-cf15-4b30-ba2e-846e0daed72d-registry-tls\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320621 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d9f5d3d-fb18-4302-8b6b-62d4b8999b65-config-volume\") pod \"collect-profiles-29517780-v65bx\" (UID: \"6d9f5d3d-fb18-4302-8b6b-62d4b8999b65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320700 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb5wr\" (UniqueName: \"kubernetes.io/projected/261ad318-57b3-48ca-8dac-e8bd53528694-kube-api-access-pb5wr\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320726 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwplk\" (UniqueName: \"kubernetes.io/projected/6d9f5d3d-fb18-4302-8b6b-62d4b8999b65-kube-api-access-rwplk\") pod \"collect-profiles-29517780-v65bx\" (UID: \"6d9f5d3d-fb18-4302-8b6b-62d4b8999b65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320960 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb4t9\" (UniqueName: \"kubernetes.io/projected/5db3acfa-ac25-413e-ac37-4a782745ab08-kube-api-access-mb4t9\") pod \"olm-operator-6b444d44fb-zpj2n\" (UID: \"5db3acfa-ac25-413e-ac37-4a782745ab08\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.320986 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91b7cb2e-1781-41da-9c90-13b40be0a2b4-service-ca-bundle\") pod \"router-default-5444994796-qh6sr\" (UID: \"91b7cb2e-1781-41da-9c90-13b40be0a2b4\") " pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.321037 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/261ad318-57b3-48ca-8dac-e8bd53528694-console-oauth-config\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.321060 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwngl\" (UniqueName: \"kubernetes.io/projected/d910fb92-a40a-4167-be6f-113f22c206b2-kube-api-access-jwngl\") pod \"route-controller-manager-6576b87f9c-rnf52\" (UID: \"d910fb92-a40a-4167-be6f-113f22c206b2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.321093 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5db3acfa-ac25-413e-ac37-4a782745ab08-srv-cert\") pod \"olm-operator-6b444d44fb-zpj2n\" (UID: \"5db3acfa-ac25-413e-ac37-4a782745ab08\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.321156 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d97a9335-cf15-4b30-ba2e-846e0daed72d-trusted-ca\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.322870 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/91b7cb2e-1781-41da-9c90-13b40be0a2b4-metrics-certs\") pod \"router-default-5444994796-qh6sr\" (UID: \"91b7cb2e-1781-41da-9c90-13b40be0a2b4\") " pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.322902 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-qjrkr\" (UID: \"fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qjrkr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.323477 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/477826a2-46cd-4145-afc9-92b69930526e-config\") pod \"kube-apiserver-operator-766d6c64bb-c5fmb\" (UID: \"477826a2-46cd-4145-afc9-92b69930526e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c5fmb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.323645 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5db3acfa-ac25-413e-ac37-4a782745ab08-profile-collector-cert\") pod \"olm-operator-6b444d44fb-zpj2n\" (UID: \"5db3acfa-ac25-413e-ac37-4a782745ab08\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.323703 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-trusted-ca-bundle\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.324472 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bbd4ae38-0244-445f-a4e8-6bffc1714d2d-serving-cert\") pod \"service-ca-operator-777779d784-hfw59\" (UID: \"bbd4ae38-0244-445f-a4e8-6bffc1714d2d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hfw59" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.324521 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/91b7cb2e-1781-41da-9c90-13b40be0a2b4-default-certificate\") pod \"router-default-5444994796-qh6sr\" (UID: \"91b7cb2e-1781-41da-9c90-13b40be0a2b4\") " pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.324542 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d910fb92-a40a-4167-be6f-113f22c206b2-serving-cert\") pod \"route-controller-manager-6576b87f9c-rnf52\" (UID: \"d910fb92-a40a-4167-be6f-113f22c206b2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.325622 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-service-ca\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.325646 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9d77fde-3517-4aea-abcb-055541763e48-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4n756\" (UID: \"e9d77fde-3517-4aea-abcb-055541763e48\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4n756" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.325704 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfxvb\" (UniqueName: \"kubernetes.io/projected/b1710cb1-bd05-4970-9b17-95cf75b9aec4-kube-api-access-sfxvb\") pod \"multus-admission-controller-857f4d67dd-x4nlj\" (UID: \"b1710cb1-bd05-4970-9b17-95cf75b9aec4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-x4nlj" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.325732 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrx55\" (UniqueName: \"kubernetes.io/projected/5897936b-fc22-4350-a0f8-c4397854da88-kube-api-access-jrx55\") pod \"migrator-59844c95c7-2gsdv\" (UID: \"5897936b-fc22-4350-a0f8-c4397854da88\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2gsdv" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.325766 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-oauth-serving-cert\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.325803 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxcc5\" (UniqueName: \"kubernetes.io/projected/bbd4ae38-0244-445f-a4e8-6bffc1714d2d-kube-api-access-kxcc5\") pod \"service-ca-operator-777779d784-hfw59\" (UID: \"bbd4ae38-0244-445f-a4e8-6bffc1714d2d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hfw59" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.325817 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftzw2\" (UniqueName: \"kubernetes.io/projected/fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d-kube-api-access-ftzw2\") pod \"machine-config-controller-84d6567774-qjrkr\" (UID: \"fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qjrkr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.325890 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.325910 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f130a8a5-24b3-4916-a0c7-b1fa5e55cad4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9pzkg\" (UID: \"f130a8a5-24b3-4916-a0c7-b1fa5e55cad4\") " pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.325953 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d97a9335-cf15-4b30-ba2e-846e0daed72d-registry-certificates\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.325967 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d910fb92-a40a-4167-be6f-113f22c206b2-config\") pod \"route-controller-manager-6576b87f9c-rnf52\" (UID: \"d910fb92-a40a-4167-be6f-113f22c206b2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.325982 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-console-config\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.325996 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brtc6\" (UniqueName: \"kubernetes.io/projected/ca75271d-7ae7-4a71-af10-732866e33d74-kube-api-access-brtc6\") pod \"control-plane-machine-set-operator-78cbb6b69f-sjcqb\" (UID: \"ca75271d-7ae7-4a71-af10-732866e33d74\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjcqb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.326021 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d9f5d3d-fb18-4302-8b6b-62d4b8999b65-secret-volume\") pod \"collect-profiles-29517780-v65bx\" (UID: \"6d9f5d3d-fb18-4302-8b6b-62d4b8999b65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.326086 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d97a9335-cf15-4b30-ba2e-846e0daed72d-bound-sa-token\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.326109 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2v69n\" (UniqueName: \"kubernetes.io/projected/91b7cb2e-1781-41da-9c90-13b40be0a2b4-kube-api-access-2v69n\") pod \"router-default-5444994796-qh6sr\" (UID: \"91b7cb2e-1781-41da-9c90-13b40be0a2b4\") " pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.326147 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/652a99d8-a666-4c63-8bc5-71953e836b01-profile-collector-cert\") pod \"catalog-operator-68c6474976-tbxbl\" (UID: \"652a99d8-a666-4c63-8bc5-71953e836b01\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.326162 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3b968f92-c773-4c84-bc7b-d00bf7a6f72b-metrics-tls\") pod \"dns-operator-744455d44c-7h7pz\" (UID: \"3b968f92-c773-4c84-bc7b-d00bf7a6f72b\") " pod="openshift-dns-operator/dns-operator-744455d44c-7h7pz" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.326188 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpr69\" (UniqueName: \"kubernetes.io/projected/d97a9335-cf15-4b30-ba2e-846e0daed72d-kube-api-access-xpr69\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.326202 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd2vw\" (UniqueName: \"kubernetes.io/projected/3b968f92-c773-4c84-bc7b-d00bf7a6f72b-kube-api-access-hd2vw\") pod \"dns-operator-744455d44c-7h7pz\" (UID: \"3b968f92-c773-4c84-bc7b-d00bf7a6f72b\") " pod="openshift-dns-operator/dns-operator-744455d44c-7h7pz" Feb 14 11:12:30 crc kubenswrapper[4904]: E0214 11:12:30.328576 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:30.828563023 +0000 UTC m=+141.641327784 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:30 crc kubenswrapper[4904]: W0214 11:12:30.352935 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb42c5242_55b6_43aa_a9bf_e368d446c3a5.slice/crio-51176d3bd0cb309dad6e16d3458d8bcc3051d7df4a3df946f6c246e8a579cdbc WatchSource:0}: Error finding container 51176d3bd0cb309dad6e16d3458d8bcc3051d7df4a3df946f6c246e8a579cdbc: Status 404 returned error can't find the container with id 51176d3bd0cb309dad6e16d3458d8bcc3051d7df4a3df946f6c246e8a579cdbc Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430170 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:30 crc kubenswrapper[4904]: E0214 11:12:30.430245 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:30.930228178 +0000 UTC m=+141.742992839 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430375 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91b7cb2e-1781-41da-9c90-13b40be0a2b4-service-ca-bundle\") pod \"router-default-5444994796-qh6sr\" (UID: \"91b7cb2e-1781-41da-9c90-13b40be0a2b4\") " pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430397 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb4t9\" (UniqueName: \"kubernetes.io/projected/5db3acfa-ac25-413e-ac37-4a782745ab08-kube-api-access-mb4t9\") pod \"olm-operator-6b444d44fb-zpj2n\" (UID: \"5db3acfa-ac25-413e-ac37-4a782745ab08\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430415 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/261ad318-57b3-48ca-8dac-e8bd53528694-console-oauth-config\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430445 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwngl\" (UniqueName: \"kubernetes.io/projected/d910fb92-a40a-4167-be6f-113f22c206b2-kube-api-access-jwngl\") pod \"route-controller-manager-6576b87f9c-rnf52\" (UID: \"d910fb92-a40a-4167-be6f-113f22c206b2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430459 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5db3acfa-ac25-413e-ac37-4a782745ab08-srv-cert\") pod \"olm-operator-6b444d44fb-zpj2n\" (UID: \"5db3acfa-ac25-413e-ac37-4a782745ab08\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430477 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2-metrics-tls\") pod \"dns-default-fjbmw\" (UID: \"ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2\") " pod="openshift-dns/dns-default-fjbmw" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430493 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d97a9335-cf15-4b30-ba2e-846e0daed72d-trusted-ca\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430527 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/91b7cb2e-1781-41da-9c90-13b40be0a2b4-metrics-certs\") pod \"router-default-5444994796-qh6sr\" (UID: \"91b7cb2e-1781-41da-9c90-13b40be0a2b4\") " pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430541 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-qjrkr\" (UID: \"fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qjrkr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430562 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/966787c0-c9ad-460e-be70-26cb6acbcf0f-mountpoint-dir\") pod \"csi-hostpathplugin-qhldd\" (UID: \"966787c0-c9ad-460e-be70-26cb6acbcf0f\") " pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430579 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/477826a2-46cd-4145-afc9-92b69930526e-config\") pod \"kube-apiserver-operator-766d6c64bb-c5fmb\" (UID: \"477826a2-46cd-4145-afc9-92b69930526e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c5fmb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430593 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5db3acfa-ac25-413e-ac37-4a782745ab08-profile-collector-cert\") pod \"olm-operator-6b444d44fb-zpj2n\" (UID: \"5db3acfa-ac25-413e-ac37-4a782745ab08\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430608 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bbd4ae38-0244-445f-a4e8-6bffc1714d2d-serving-cert\") pod \"service-ca-operator-777779d784-hfw59\" (UID: \"bbd4ae38-0244-445f-a4e8-6bffc1714d2d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hfw59" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430629 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-trusted-ca-bundle\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430643 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/91b7cb2e-1781-41da-9c90-13b40be0a2b4-default-certificate\") pod \"router-default-5444994796-qh6sr\" (UID: \"91b7cb2e-1781-41da-9c90-13b40be0a2b4\") " pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430659 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/966787c0-c9ad-460e-be70-26cb6acbcf0f-csi-data-dir\") pod \"csi-hostpathplugin-qhldd\" (UID: \"966787c0-c9ad-460e-be70-26cb6acbcf0f\") " pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430675 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d910fb92-a40a-4167-be6f-113f22c206b2-serving-cert\") pod \"route-controller-manager-6576b87f9c-rnf52\" (UID: \"d910fb92-a40a-4167-be6f-113f22c206b2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430701 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/966787c0-c9ad-460e-be70-26cb6acbcf0f-socket-dir\") pod \"csi-hostpathplugin-qhldd\" (UID: \"966787c0-c9ad-460e-be70-26cb6acbcf0f\") " pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430725 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-service-ca\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430745 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9d77fde-3517-4aea-abcb-055541763e48-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4n756\" (UID: \"e9d77fde-3517-4aea-abcb-055541763e48\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4n756" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430761 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfxvb\" (UniqueName: \"kubernetes.io/projected/b1710cb1-bd05-4970-9b17-95cf75b9aec4-kube-api-access-sfxvb\") pod \"multus-admission-controller-857f4d67dd-x4nlj\" (UID: \"b1710cb1-bd05-4970-9b17-95cf75b9aec4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-x4nlj" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430790 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c218d2bc-6512-4c03-b39e-05c801dac4b1-certs\") pod \"machine-config-server-6wnjb\" (UID: \"c218d2bc-6512-4c03-b39e-05c801dac4b1\") " pod="openshift-machine-config-operator/machine-config-server-6wnjb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430807 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrx55\" (UniqueName: \"kubernetes.io/projected/5897936b-fc22-4350-a0f8-c4397854da88-kube-api-access-jrx55\") pod \"migrator-59844c95c7-2gsdv\" (UID: \"5897936b-fc22-4350-a0f8-c4397854da88\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2gsdv" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430821 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b6796678-081b-4588-88d2-1cf0ef013ba4-cert\") pod \"ingress-canary-2q49s\" (UID: \"b6796678-081b-4588-88d2-1cf0ef013ba4\") " pod="openshift-ingress-canary/ingress-canary-2q49s" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430858 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-oauth-serving-cert\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430912 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxcc5\" (UniqueName: \"kubernetes.io/projected/bbd4ae38-0244-445f-a4e8-6bffc1714d2d-kube-api-access-kxcc5\") pod \"service-ca-operator-777779d784-hfw59\" (UID: \"bbd4ae38-0244-445f-a4e8-6bffc1714d2d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hfw59" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430929 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftzw2\" (UniqueName: \"kubernetes.io/projected/fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d-kube-api-access-ftzw2\") pod \"machine-config-controller-84d6567774-qjrkr\" (UID: \"fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qjrkr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.430948 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431069 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f130a8a5-24b3-4916-a0c7-b1fa5e55cad4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9pzkg\" (UID: \"f130a8a5-24b3-4916-a0c7-b1fa5e55cad4\") " pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431097 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr6k2\" (UniqueName: \"kubernetes.io/projected/ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2-kube-api-access-jr6k2\") pod \"dns-default-fjbmw\" (UID: \"ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2\") " pod="openshift-dns/dns-default-fjbmw" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431114 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d97a9335-cf15-4b30-ba2e-846e0daed72d-registry-certificates\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431131 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d910fb92-a40a-4167-be6f-113f22c206b2-config\") pod \"route-controller-manager-6576b87f9c-rnf52\" (UID: \"d910fb92-a40a-4167-be6f-113f22c206b2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431145 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-console-config\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431159 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brtc6\" (UniqueName: \"kubernetes.io/projected/ca75271d-7ae7-4a71-af10-732866e33d74-kube-api-access-brtc6\") pod \"control-plane-machine-set-operator-78cbb6b69f-sjcqb\" (UID: \"ca75271d-7ae7-4a71-af10-732866e33d74\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjcqb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431183 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d9f5d3d-fb18-4302-8b6b-62d4b8999b65-secret-volume\") pod \"collect-profiles-29517780-v65bx\" (UID: \"6d9f5d3d-fb18-4302-8b6b-62d4b8999b65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431200 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d97a9335-cf15-4b30-ba2e-846e0daed72d-bound-sa-token\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431222 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2v69n\" (UniqueName: \"kubernetes.io/projected/91b7cb2e-1781-41da-9c90-13b40be0a2b4-kube-api-access-2v69n\") pod \"router-default-5444994796-qh6sr\" (UID: \"91b7cb2e-1781-41da-9c90-13b40be0a2b4\") " pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431238 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/652a99d8-a666-4c63-8bc5-71953e836b01-profile-collector-cert\") pod \"catalog-operator-68c6474976-tbxbl\" (UID: \"652a99d8-a666-4c63-8bc5-71953e836b01\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431252 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3b968f92-c773-4c84-bc7b-d00bf7a6f72b-metrics-tls\") pod \"dns-operator-744455d44c-7h7pz\" (UID: \"3b968f92-c773-4c84-bc7b-d00bf7a6f72b\") " pod="openshift-dns-operator/dns-operator-744455d44c-7h7pz" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431267 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqrfp\" (UniqueName: \"kubernetes.io/projected/966787c0-c9ad-460e-be70-26cb6acbcf0f-kube-api-access-fqrfp\") pod \"csi-hostpathplugin-qhldd\" (UID: \"966787c0-c9ad-460e-be70-26cb6acbcf0f\") " pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431283 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpr69\" (UniqueName: \"kubernetes.io/projected/d97a9335-cf15-4b30-ba2e-846e0daed72d-kube-api-access-xpr69\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431299 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd2vw\" (UniqueName: \"kubernetes.io/projected/3b968f92-c773-4c84-bc7b-d00bf7a6f72b-kube-api-access-hd2vw\") pod \"dns-operator-744455d44c-7h7pz\" (UID: \"3b968f92-c773-4c84-bc7b-d00bf7a6f72b\") " pod="openshift-dns-operator/dns-operator-744455d44c-7h7pz" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431323 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/477826a2-46cd-4145-afc9-92b69930526e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-c5fmb\" (UID: \"477826a2-46cd-4145-afc9-92b69930526e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c5fmb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431348 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/966787c0-c9ad-460e-be70-26cb6acbcf0f-plugins-dir\") pod \"csi-hostpathplugin-qhldd\" (UID: \"966787c0-c9ad-460e-be70-26cb6acbcf0f\") " pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431377 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2-config-volume\") pod \"dns-default-fjbmw\" (UID: \"ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2\") " pod="openshift-dns/dns-default-fjbmw" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431392 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d97a9335-cf15-4b30-ba2e-846e0daed72d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431407 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/261ad318-57b3-48ca-8dac-e8bd53528694-console-serving-cert\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431441 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d77fde-3517-4aea-abcb-055541763e48-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4n756\" (UID: \"e9d77fde-3517-4aea-abcb-055541763e48\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4n756" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431460 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d-proxy-tls\") pod \"machine-config-controller-84d6567774-qjrkr\" (UID: \"fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qjrkr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431477 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/966787c0-c9ad-460e-be70-26cb6acbcf0f-registration-dir\") pod \"csi-hostpathplugin-qhldd\" (UID: \"966787c0-c9ad-460e-be70-26cb6acbcf0f\") " pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431494 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hklmr\" (UniqueName: \"kubernetes.io/projected/c218d2bc-6512-4c03-b39e-05c801dac4b1-kube-api-access-hklmr\") pod \"machine-config-server-6wnjb\" (UID: \"c218d2bc-6512-4c03-b39e-05c801dac4b1\") " pod="openshift-machine-config-operator/machine-config-server-6wnjb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431520 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/477826a2-46cd-4145-afc9-92b69930526e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-c5fmb\" (UID: \"477826a2-46cd-4145-afc9-92b69930526e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c5fmb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431534 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbd4ae38-0244-445f-a4e8-6bffc1714d2d-config\") pod \"service-ca-operator-777779d784-hfw59\" (UID: \"bbd4ae38-0244-445f-a4e8-6bffc1714d2d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hfw59" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431553 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/652a99d8-a666-4c63-8bc5-71953e836b01-srv-cert\") pod \"catalog-operator-68c6474976-tbxbl\" (UID: \"652a99d8-a666-4c63-8bc5-71953e836b01\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431569 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f130a8a5-24b3-4916-a0c7-b1fa5e55cad4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9pzkg\" (UID: \"f130a8a5-24b3-4916-a0c7-b1fa5e55cad4\") " pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431584 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfn4h\" (UniqueName: \"kubernetes.io/projected/652a99d8-a666-4c63-8bc5-71953e836b01-kube-api-access-pfn4h\") pod \"catalog-operator-68c6474976-tbxbl\" (UID: \"652a99d8-a666-4c63-8bc5-71953e836b01\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431598 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/91b7cb2e-1781-41da-9c90-13b40be0a2b4-stats-auth\") pod \"router-default-5444994796-qh6sr\" (UID: \"91b7cb2e-1781-41da-9c90-13b40be0a2b4\") " pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431623 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b1710cb1-bd05-4970-9b17-95cf75b9aec4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-x4nlj\" (UID: \"b1710cb1-bd05-4970-9b17-95cf75b9aec4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-x4nlj" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431648 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d97a9335-cf15-4b30-ba2e-846e0daed72d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431681 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d910fb92-a40a-4167-be6f-113f22c206b2-client-ca\") pod \"route-controller-manager-6576b87f9c-rnf52\" (UID: \"d910fb92-a40a-4167-be6f-113f22c206b2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431695 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzzg8\" (UniqueName: \"kubernetes.io/projected/f130a8a5-24b3-4916-a0c7-b1fa5e55cad4-kube-api-access-pzzg8\") pod \"marketplace-operator-79b997595-9pzkg\" (UID: \"f130a8a5-24b3-4916-a0c7-b1fa5e55cad4\") " pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431710 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ca75271d-7ae7-4a71-af10-732866e33d74-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sjcqb\" (UID: \"ca75271d-7ae7-4a71-af10-732866e33d74\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjcqb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431734 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d77fde-3517-4aea-abcb-055541763e48-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4n756\" (UID: \"e9d77fde-3517-4aea-abcb-055541763e48\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4n756" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431749 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d97a9335-cf15-4b30-ba2e-846e0daed72d-registry-tls\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431763 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d9f5d3d-fb18-4302-8b6b-62d4b8999b65-config-volume\") pod \"collect-profiles-29517780-v65bx\" (UID: \"6d9f5d3d-fb18-4302-8b6b-62d4b8999b65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431797 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb5wr\" (UniqueName: \"kubernetes.io/projected/261ad318-57b3-48ca-8dac-e8bd53528694-kube-api-access-pb5wr\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431815 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c218d2bc-6512-4c03-b39e-05c801dac4b1-node-bootstrap-token\") pod \"machine-config-server-6wnjb\" (UID: \"c218d2bc-6512-4c03-b39e-05c801dac4b1\") " pod="openshift-machine-config-operator/machine-config-server-6wnjb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431858 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwplk\" (UniqueName: \"kubernetes.io/projected/6d9f5d3d-fb18-4302-8b6b-62d4b8999b65-kube-api-access-rwplk\") pod \"collect-profiles-29517780-v65bx\" (UID: \"6d9f5d3d-fb18-4302-8b6b-62d4b8999b65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.431873 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c899\" (UniqueName: \"kubernetes.io/projected/b6796678-081b-4588-88d2-1cf0ef013ba4-kube-api-access-4c899\") pod \"ingress-canary-2q49s\" (UID: \"b6796678-081b-4588-88d2-1cf0ef013ba4\") " pod="openshift-ingress-canary/ingress-canary-2q49s" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.436619 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-49jn7" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.437385 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91b7cb2e-1781-41da-9c90-13b40be0a2b4-service-ca-bundle\") pod \"router-default-5444994796-qh6sr\" (UID: \"91b7cb2e-1781-41da-9c90-13b40be0a2b4\") " pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.437760 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/477826a2-46cd-4145-afc9-92b69930526e-config\") pod \"kube-apiserver-operator-766d6c64bb-c5fmb\" (UID: \"477826a2-46cd-4145-afc9-92b69930526e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c5fmb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.438201 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-qjrkr\" (UID: \"fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qjrkr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.442480 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/477826a2-46cd-4145-afc9-92b69930526e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-c5fmb\" (UID: \"477826a2-46cd-4145-afc9-92b69930526e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c5fmb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.443472 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/91b7cb2e-1781-41da-9c90-13b40be0a2b4-metrics-certs\") pod \"router-default-5444994796-qh6sr\" (UID: \"91b7cb2e-1781-41da-9c90-13b40be0a2b4\") " pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.444432 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d77fde-3517-4aea-abcb-055541763e48-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4n756\" (UID: \"e9d77fde-3517-4aea-abcb-055541763e48\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4n756" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.444800 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d97a9335-cf15-4b30-ba2e-846e0daed72d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.445360 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-oauth-serving-cert\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.459795 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/261ad318-57b3-48ca-8dac-e8bd53528694-console-oauth-config\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.467306 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d97a9335-cf15-4b30-ba2e-846e0daed72d-registry-certificates\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.468050 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbd4ae38-0244-445f-a4e8-6bffc1714d2d-config\") pod \"service-ca-operator-777779d784-hfw59\" (UID: \"bbd4ae38-0244-445f-a4e8-6bffc1714d2d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hfw59" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.471137 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f130a8a5-24b3-4916-a0c7-b1fa5e55cad4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9pzkg\" (UID: \"f130a8a5-24b3-4916-a0c7-b1fa5e55cad4\") " pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.481397 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d910fb92-a40a-4167-be6f-113f22c206b2-config\") pod \"route-controller-manager-6576b87f9c-rnf52\" (UID: \"d910fb92-a40a-4167-be6f-113f22c206b2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.482657 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ql7rx"] Feb 14 11:12:30 crc kubenswrapper[4904]: E0214 11:12:30.483401 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:30.983387432 +0000 UTC m=+141.796152093 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.484705 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5db3acfa-ac25-413e-ac37-4a782745ab08-profile-collector-cert\") pod \"olm-operator-6b444d44fb-zpj2n\" (UID: \"5db3acfa-ac25-413e-ac37-4a782745ab08\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.485993 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.486733 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d910fb92-a40a-4167-be6f-113f22c206b2-client-ca\") pod \"route-controller-manager-6576b87f9c-rnf52\" (UID: \"d910fb92-a40a-4167-be6f-113f22c206b2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.487019 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5db3acfa-ac25-413e-ac37-4a782745ab08-srv-cert\") pod \"olm-operator-6b444d44fb-zpj2n\" (UID: \"5db3acfa-ac25-413e-ac37-4a782745ab08\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.490568 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/261ad318-57b3-48ca-8dac-e8bd53528694-console-serving-cert\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.490972 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-trusted-ca-bundle\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.491019 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d97a9335-cf15-4b30-ba2e-846e0daed72d-trusted-ca\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.491308 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-console-config\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.491779 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d9f5d3d-fb18-4302-8b6b-62d4b8999b65-secret-volume\") pod \"collect-profiles-29517780-v65bx\" (UID: \"6d9f5d3d-fb18-4302-8b6b-62d4b8999b65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.493808 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d9f5d3d-fb18-4302-8b6b-62d4b8999b65-config-volume\") pod \"collect-profiles-29517780-v65bx\" (UID: \"6d9f5d3d-fb18-4302-8b6b-62d4b8999b65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.494041 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-service-ca\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.498355 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/91b7cb2e-1781-41da-9c90-13b40be0a2b4-default-certificate\") pod \"router-default-5444994796-qh6sr\" (UID: \"91b7cb2e-1781-41da-9c90-13b40be0a2b4\") " pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.499177 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/652a99d8-a666-4c63-8bc5-71953e836b01-srv-cert\") pod \"catalog-operator-68c6474976-tbxbl\" (UID: \"652a99d8-a666-4c63-8bc5-71953e836b01\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.499770 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b1710cb1-bd05-4970-9b17-95cf75b9aec4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-x4nlj\" (UID: \"b1710cb1-bd05-4970-9b17-95cf75b9aec4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-x4nlj" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.501544 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d97a9335-cf15-4b30-ba2e-846e0daed72d-registry-tls\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.504519 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d97a9335-cf15-4b30-ba2e-846e0daed72d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.509120 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp"] Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.509172 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-bzps5"] Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.509298 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/91b7cb2e-1781-41da-9c90-13b40be0a2b4-stats-auth\") pod \"router-default-5444994796-qh6sr\" (UID: \"91b7cb2e-1781-41da-9c90-13b40be0a2b4\") " pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.516388 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx"] Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.518466 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpr69\" (UniqueName: \"kubernetes.io/projected/d97a9335-cf15-4b30-ba2e-846e0daed72d-kube-api-access-xpr69\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.519205 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3b968f92-c773-4c84-bc7b-d00bf7a6f72b-metrics-tls\") pod \"dns-operator-744455d44c-7h7pz\" (UID: \"3b968f92-c773-4c84-bc7b-d00bf7a6f72b\") " pod="openshift-dns-operator/dns-operator-744455d44c-7h7pz" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.519387 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d910fb92-a40a-4167-be6f-113f22c206b2-serving-cert\") pod \"route-controller-manager-6576b87f9c-rnf52\" (UID: \"d910fb92-a40a-4167-be6f-113f22c206b2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.521327 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d-proxy-tls\") pod \"machine-config-controller-84d6567774-qjrkr\" (UID: \"fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qjrkr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.521909 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ca75271d-7ae7-4a71-af10-732866e33d74-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sjcqb\" (UID: \"ca75271d-7ae7-4a71-af10-732866e33d74\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjcqb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.526585 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bbd4ae38-0244-445f-a4e8-6bffc1714d2d-serving-cert\") pod \"service-ca-operator-777779d784-hfw59\" (UID: \"bbd4ae38-0244-445f-a4e8-6bffc1714d2d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hfw59" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.526705 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d77fde-3517-4aea-abcb-055541763e48-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4n756\" (UID: \"e9d77fde-3517-4aea-abcb-055541763e48\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4n756" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.526898 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f130a8a5-24b3-4916-a0c7-b1fa5e55cad4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9pzkg\" (UID: \"f130a8a5-24b3-4916-a0c7-b1fa5e55cad4\") " pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.527968 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/652a99d8-a666-4c63-8bc5-71953e836b01-profile-collector-cert\") pod \"catalog-operator-68c6474976-tbxbl\" (UID: \"652a99d8-a666-4c63-8bc5-71953e836b01\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.536476 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.537732 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2-config-volume\") pod \"dns-default-fjbmw\" (UID: \"ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2\") " pod="openshift-dns/dns-default-fjbmw" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.537780 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/966787c0-c9ad-460e-be70-26cb6acbcf0f-registration-dir\") pod \"csi-hostpathplugin-qhldd\" (UID: \"966787c0-c9ad-460e-be70-26cb6acbcf0f\") " pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.537803 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hklmr\" (UniqueName: \"kubernetes.io/projected/c218d2bc-6512-4c03-b39e-05c801dac4b1-kube-api-access-hklmr\") pod \"machine-config-server-6wnjb\" (UID: \"c218d2bc-6512-4c03-b39e-05c801dac4b1\") " pod="openshift-machine-config-operator/machine-config-server-6wnjb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.537907 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c218d2bc-6512-4c03-b39e-05c801dac4b1-node-bootstrap-token\") pod \"machine-config-server-6wnjb\" (UID: \"c218d2bc-6512-4c03-b39e-05c801dac4b1\") " pod="openshift-machine-config-operator/machine-config-server-6wnjb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.537932 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c899\" (UniqueName: \"kubernetes.io/projected/b6796678-081b-4588-88d2-1cf0ef013ba4-kube-api-access-4c899\") pod \"ingress-canary-2q49s\" (UID: \"b6796678-081b-4588-88d2-1cf0ef013ba4\") " pod="openshift-ingress-canary/ingress-canary-2q49s" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.537974 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2-metrics-tls\") pod \"dns-default-fjbmw\" (UID: \"ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2\") " pod="openshift-dns/dns-default-fjbmw" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.538001 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/966787c0-c9ad-460e-be70-26cb6acbcf0f-mountpoint-dir\") pod \"csi-hostpathplugin-qhldd\" (UID: \"966787c0-c9ad-460e-be70-26cb6acbcf0f\") " pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.538034 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/966787c0-c9ad-460e-be70-26cb6acbcf0f-csi-data-dir\") pod \"csi-hostpathplugin-qhldd\" (UID: \"966787c0-c9ad-460e-be70-26cb6acbcf0f\") " pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.538060 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/966787c0-c9ad-460e-be70-26cb6acbcf0f-socket-dir\") pod \"csi-hostpathplugin-qhldd\" (UID: \"966787c0-c9ad-460e-be70-26cb6acbcf0f\") " pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.538093 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c218d2bc-6512-4c03-b39e-05c801dac4b1-certs\") pod \"machine-config-server-6wnjb\" (UID: \"c218d2bc-6512-4c03-b39e-05c801dac4b1\") " pod="openshift-machine-config-operator/machine-config-server-6wnjb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.538119 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b6796678-081b-4588-88d2-1cf0ef013ba4-cert\") pod \"ingress-canary-2q49s\" (UID: \"b6796678-081b-4588-88d2-1cf0ef013ba4\") " pod="openshift-ingress-canary/ingress-canary-2q49s" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.538172 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr6k2\" (UniqueName: \"kubernetes.io/projected/ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2-kube-api-access-jr6k2\") pod \"dns-default-fjbmw\" (UID: \"ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2\") " pod="openshift-dns/dns-default-fjbmw" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.538250 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqrfp\" (UniqueName: \"kubernetes.io/projected/966787c0-c9ad-460e-be70-26cb6acbcf0f-kube-api-access-fqrfp\") pod \"csi-hostpathplugin-qhldd\" (UID: \"966787c0-c9ad-460e-be70-26cb6acbcf0f\") " pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.538285 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/966787c0-c9ad-460e-be70-26cb6acbcf0f-plugins-dir\") pod \"csi-hostpathplugin-qhldd\" (UID: \"966787c0-c9ad-460e-be70-26cb6acbcf0f\") " pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.538578 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/966787c0-c9ad-460e-be70-26cb6acbcf0f-plugins-dir\") pod \"csi-hostpathplugin-qhldd\" (UID: \"966787c0-c9ad-460e-be70-26cb6acbcf0f\") " pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:30 crc kubenswrapper[4904]: E0214 11:12:30.538660 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:31.038640311 +0000 UTC m=+141.851404972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.539223 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2-config-volume\") pod \"dns-default-fjbmw\" (UID: \"ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2\") " pod="openshift-dns/dns-default-fjbmw" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.539291 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/966787c0-c9ad-460e-be70-26cb6acbcf0f-registration-dir\") pod \"csi-hostpathplugin-qhldd\" (UID: \"966787c0-c9ad-460e-be70-26cb6acbcf0f\") " pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.543173 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb4t9\" (UniqueName: \"kubernetes.io/projected/5db3acfa-ac25-413e-ac37-4a782745ab08-kube-api-access-mb4t9\") pod \"olm-operator-6b444d44fb-zpj2n\" (UID: \"5db3acfa-ac25-413e-ac37-4a782745ab08\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.543999 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/966787c0-c9ad-460e-be70-26cb6acbcf0f-socket-dir\") pod \"csi-hostpathplugin-qhldd\" (UID: \"966787c0-c9ad-460e-be70-26cb6acbcf0f\") " pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.546207 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/966787c0-c9ad-460e-be70-26cb6acbcf0f-mountpoint-dir\") pod \"csi-hostpathplugin-qhldd\" (UID: \"966787c0-c9ad-460e-be70-26cb6acbcf0f\") " pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.550065 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw"] Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.553087 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-d948p"] Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.555723 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd2vw\" (UniqueName: \"kubernetes.io/projected/3b968f92-c773-4c84-bc7b-d00bf7a6f72b-kube-api-access-hd2vw\") pod \"dns-operator-744455d44c-7h7pz\" (UID: \"3b968f92-c773-4c84-bc7b-d00bf7a6f72b\") " pod="openshift-dns-operator/dns-operator-744455d44c-7h7pz" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.558757 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/966787c0-c9ad-460e-be70-26cb6acbcf0f-csi-data-dir\") pod \"csi-hostpathplugin-qhldd\" (UID: \"966787c0-c9ad-460e-be70-26cb6acbcf0f\") " pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.562758 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c218d2bc-6512-4c03-b39e-05c801dac4b1-node-bootstrap-token\") pod \"machine-config-server-6wnjb\" (UID: \"c218d2bc-6512-4c03-b39e-05c801dac4b1\") " pod="openshift-machine-config-operator/machine-config-server-6wnjb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.564661 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwngl\" (UniqueName: \"kubernetes.io/projected/d910fb92-a40a-4167-be6f-113f22c206b2-kube-api-access-jwngl\") pod \"route-controller-manager-6576b87f9c-rnf52\" (UID: \"d910fb92-a40a-4167-be6f-113f22c206b2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.566189 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c218d2bc-6512-4c03-b39e-05c801dac4b1-certs\") pod \"machine-config-server-6wnjb\" (UID: \"c218d2bc-6512-4c03-b39e-05c801dac4b1\") " pod="openshift-machine-config-operator/machine-config-server-6wnjb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.577851 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2-metrics-tls\") pod \"dns-default-fjbmw\" (UID: \"ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2\") " pod="openshift-dns/dns-default-fjbmw" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.578297 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b6796678-081b-4588-88d2-1cf0ef013ba4-cert\") pod \"ingress-canary-2q49s\" (UID: \"b6796678-081b-4588-88d2-1cf0ef013ba4\") " pod="openshift-ingress-canary/ingress-canary-2q49s" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.581682 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jq824"] Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.581997 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brtc6\" (UniqueName: \"kubernetes.io/projected/ca75271d-7ae7-4a71-af10-732866e33d74-kube-api-access-brtc6\") pod \"control-plane-machine-set-operator-78cbb6b69f-sjcqb\" (UID: \"ca75271d-7ae7-4a71-af10-732866e33d74\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjcqb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.583240 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9d77fde-3517-4aea-abcb-055541763e48-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4n756\" (UID: \"e9d77fde-3517-4aea-abcb-055541763e48\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4n756" Feb 14 11:12:30 crc kubenswrapper[4904]: W0214 11:12:30.586315 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3cb9bb68_546e_4e52_9a15_5aeee129a5b4.slice/crio-fc9e9027476ba75438f2538fa79e043110011e65da9a19add1cd49f81eabaabc WatchSource:0}: Error finding container fc9e9027476ba75438f2538fa79e043110011e65da9a19add1cd49f81eabaabc: Status 404 returned error can't find the container with id fc9e9027476ba75438f2538fa79e043110011e65da9a19add1cd49f81eabaabc Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.603820 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-f95s6"] Feb 14 11:12:30 crc kubenswrapper[4904]: W0214 11:12:30.606253 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod821984e5_b740_4332_9c06_1ccefda9a9fb.slice/crio-64d1fa32a191863bdcb8ccedb51937e19bceda7c7a6c694ea60e6e7cebbd120c WatchSource:0}: Error finding container 64d1fa32a191863bdcb8ccedb51937e19bceda7c7a6c694ea60e6e7cebbd120c: Status 404 returned error can't find the container with id 64d1fa32a191863bdcb8ccedb51937e19bceda7c7a6c694ea60e6e7cebbd120c Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.611477 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzzg8\" (UniqueName: \"kubernetes.io/projected/f130a8a5-24b3-4916-a0c7-b1fa5e55cad4-kube-api-access-pzzg8\") pod \"marketplace-operator-79b997595-9pzkg\" (UID: \"f130a8a5-24b3-4916-a0c7-b1fa5e55cad4\") " pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.615507 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-d9d4k"] Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.627453 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/477826a2-46cd-4145-afc9-92b69930526e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-c5fmb\" (UID: \"477826a2-46cd-4145-afc9-92b69930526e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c5fmb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.639170 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfxvb\" (UniqueName: \"kubernetes.io/projected/b1710cb1-bd05-4970-9b17-95cf75b9aec4-kube-api-access-sfxvb\") pod \"multus-admission-controller-857f4d67dd-x4nlj\" (UID: \"b1710cb1-bd05-4970-9b17-95cf75b9aec4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-x4nlj" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.640973 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: E0214 11:12:30.641535 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:31.141508097 +0000 UTC m=+141.954272758 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.683382 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrx55\" (UniqueName: \"kubernetes.io/projected/5897936b-fc22-4350-a0f8-c4397854da88-kube-api-access-jrx55\") pod \"migrator-59844c95c7-2gsdv\" (UID: \"5897936b-fc22-4350-a0f8-c4397854da88\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2gsdv" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.698778 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfn4h\" (UniqueName: \"kubernetes.io/projected/652a99d8-a666-4c63-8bc5-71953e836b01-kube-api-access-pfn4h\") pod \"catalog-operator-68c6474976-tbxbl\" (UID: \"652a99d8-a666-4c63-8bc5-71953e836b01\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.699081 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.700776 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-7h7pz" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.727594 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxcc5\" (UniqueName: \"kubernetes.io/projected/bbd4ae38-0244-445f-a4e8-6bffc1714d2d-kube-api-access-kxcc5\") pod \"service-ca-operator-777779d784-hfw59\" (UID: \"bbd4ae38-0244-445f-a4e8-6bffc1714d2d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hfw59" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.727824 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.728224 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-x4nlj" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.736166 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2gsdv" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.738020 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftzw2\" (UniqueName: \"kubernetes.io/projected/fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d-kube-api-access-ftzw2\") pod \"machine-config-controller-84d6567774-qjrkr\" (UID: \"fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qjrkr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.742577 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.742972 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c5fmb" Feb 14 11:12:30 crc kubenswrapper[4904]: E0214 11:12:30.743091 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:31.24306982 +0000 UTC m=+142.055834481 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.744725 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7t272"] Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.749486 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4n756" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.755312 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d97a9335-cf15-4b30-ba2e-846e0daed72d-bound-sa-token\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.760379 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2v69n\" (UniqueName: \"kubernetes.io/projected/91b7cb2e-1781-41da-9c90-13b40be0a2b4-kube-api-access-2v69n\") pod \"router-default-5444994796-qh6sr\" (UID: \"91b7cb2e-1781-41da-9c90-13b40be0a2b4\") " pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.763806 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjcqb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.790488 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwplk\" (UniqueName: \"kubernetes.io/projected/6d9f5d3d-fb18-4302-8b6b-62d4b8999b65-kube-api-access-rwplk\") pod \"collect-profiles-29517780-v65bx\" (UID: \"6d9f5d3d-fb18-4302-8b6b-62d4b8999b65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.791446 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.797509 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hfw59" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.804302 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.808859 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" event={"ID":"e3253b15-b01e-4c6b-91df-578ddb39b234","Type":"ContainerStarted","Data":"6c4f59e5603e9919f48c9f3ed4a03fd3907b1c72af3827e9dec3cedfc03ba1a0"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.811918 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb5wr\" (UniqueName: \"kubernetes.io/projected/261ad318-57b3-48ca-8dac-e8bd53528694-kube-api-access-pb5wr\") pod \"console-f9d7485db-7c6cp\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.817232 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.828407 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.831365 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-hgggh" event={"ID":"800acf17-404a-4dc3-a25f-66e779d292fd","Type":"ContainerStarted","Data":"95c6931961773b209b4c1a5ba15c3914920815d45f69163a7f90afa27b48f094"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.831400 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-hgggh" event={"ID":"800acf17-404a-4dc3-a25f-66e779d292fd","Type":"ContainerStarted","Data":"ca6c964457e5cf34557e8fc62067ebb28a0010ba2883828819e8c9aca67dfd33"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.831409 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-hgggh" event={"ID":"800acf17-404a-4dc3-a25f-66e779d292fd","Type":"ContainerStarted","Data":"024d65ec4477a308e7bbc778845f4279fb89dd29e2423ac6b4324438a11b8f22"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.832439 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hklmr\" (UniqueName: \"kubernetes.io/projected/c218d2bc-6512-4c03-b39e-05c801dac4b1-kube-api-access-hklmr\") pod \"machine-config-server-6wnjb\" (UID: \"c218d2bc-6512-4c03-b39e-05c801dac4b1\") " pod="openshift-machine-config-operator/machine-config-server-6wnjb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.842422 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c899\" (UniqueName: \"kubernetes.io/projected/b6796678-081b-4588-88d2-1cf0ef013ba4-kube-api-access-4c899\") pod \"ingress-canary-2q49s\" (UID: \"b6796678-081b-4588-88d2-1cf0ef013ba4\") " pod="openshift-ingress-canary/ingress-canary-2q49s" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.842792 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-6wnjb" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.843130 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2q49s" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.843899 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" event={"ID":"5a537ec4-ab86-4bc9-81d2-3870472b767a","Type":"ContainerStarted","Data":"9969e557a55d34107e04e866c201d4d8d5e8d7eb9f45fd44bde1c7338015a20e"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.843930 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" event={"ID":"5a537ec4-ab86-4bc9-81d2-3870472b767a","Type":"ContainerStarted","Data":"d84f717518ef133711c1a0c3b66797a4af5047c5dfa24914b8f74a29d8e4d9cd"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.843943 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.844314 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:30 crc kubenswrapper[4904]: E0214 11:12:30.845360 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:31.345348851 +0000 UTC m=+142.158113512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.858236 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" event={"ID":"a7f9f078-ff94-4adc-be45-927915d6e55f","Type":"ContainerStarted","Data":"e3cc7ed5471400341b533d0b668591497391cf39c33937ab2a0a084e72f76f18"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.859326 4904 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-5lwz5 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.859357 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" podUID="5a537ec4-ab86-4bc9-81d2-3870472b767a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.862639 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fqtd8"] Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.862870 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z"] Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.867508 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr6k2\" (UniqueName: \"kubernetes.io/projected/ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2-kube-api-access-jr6k2\") pod \"dns-default-fjbmw\" (UID: \"ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2\") " pod="openshift-dns/dns-default-fjbmw" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.873108 4904 generic.go:334] "Generic (PLEG): container finished" podID="80b9d666-fa50-4d49-8c3f-efbd9d1247a9" containerID="eb291fb5ba59620e8f7eec68aae7ef3cf8bfcbbb5f767d2c0921baaed2a6a3d1" exitCode=0 Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.873189 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" event={"ID":"80b9d666-fa50-4d49-8c3f-efbd9d1247a9","Type":"ContainerDied","Data":"eb291fb5ba59620e8f7eec68aae7ef3cf8bfcbbb5f767d2c0921baaed2a6a3d1"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.873215 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" event={"ID":"80b9d666-fa50-4d49-8c3f-efbd9d1247a9","Type":"ContainerStarted","Data":"31d99f22387f07606a335091e279d2d6fe37149f319a8ccfaa143d0f7394f21f"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.874608 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-fjbmw" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.875492 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bzps5" event={"ID":"3cb9bb68-546e-4e52-9a15-5aeee129a5b4","Type":"ContainerStarted","Data":"fc9e9027476ba75438f2538fa79e043110011e65da9a19add1cd49f81eabaabc"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.876296 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-d948p" event={"ID":"30e3e22f-c17f-4efa-b5cd-33406a09def1","Type":"ContainerStarted","Data":"083240b0ddb09d415da530ab76b257b9455138eec77417c8364b823159399208"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.877419 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ql7rx" event={"ID":"2201ed6d-9862-43d5-98ed-949539d6d0ae","Type":"ContainerStarted","Data":"3b7ec240b910b2f5ef108609dd0d9bedb180f7bffb90a316bd02c0c69f7a5b02"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.887424 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-f95s6" event={"ID":"7c2c0ca2-4182-4c24-8710-c3b632e768f9","Type":"ContainerStarted","Data":"e6ddf1871e492142dd74c86129154f94c5fc143d3517f7ef3e9f01c721e83aeb"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.890190 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqrfp\" (UniqueName: \"kubernetes.io/projected/966787c0-c9ad-460e-be70-26cb6acbcf0f-kube-api-access-fqrfp\") pod \"csi-hostpathplugin-qhldd\" (UID: \"966787c0-c9ad-460e-be70-26cb6acbcf0f\") " pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.915061 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-49jn7"] Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.918474 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w"] Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.919945 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" event={"ID":"821984e5-b740-4332-9c06-1ccefda9a9fb","Type":"ContainerStarted","Data":"64d1fa32a191863bdcb8ccedb51937e19bceda7c7a6c694ea60e6e7cebbd120c"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.932308 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" event={"ID":"345bb34c-a9b2-4e88-94ce-3187b8933057","Type":"ContainerStarted","Data":"317457372c4ef32b4b7b9924045deecacbcc33f5c12d3d62f9cddcf87e7c69a1"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.943383 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4" event={"ID":"8b89659e-58f4-4c11-9839-2be0bdcd6e5e","Type":"ContainerStarted","Data":"942338b7b910aa6743149c871078a270c178c1074feb595dd6b8e75df9d79027"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.943423 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4" event={"ID":"8b89659e-58f4-4c11-9839-2be0bdcd6e5e","Type":"ContainerStarted","Data":"676dd5d6d860710ec48643f22f42c09e491b05a7da9cdb0baddc346cfb9b487d"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.944346 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g25mq" event={"ID":"fdb5ad89-a263-4f1f-bbe5-42bd5b91ea25","Type":"ContainerStarted","Data":"5b3adac780cd9eca62560d6cd5fec7c117b7ba650cbc5caf280850f3b9bdc981"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.948458 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:30 crc kubenswrapper[4904]: E0214 11:12:30.949498 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:31.44947385 +0000 UTC m=+142.262238531 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.953873 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" event={"ID":"de09ed81-3849-460e-93f2-fb0f5d72a4b9","Type":"ContainerStarted","Data":"7c74402bf539b13370c5d9351c1d4ffdf8c42192239bf2fecb036b70c8776bcb"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.965484 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qsppc" event={"ID":"b42c5242-55b6-43aa-a9bf-e368d446c3a5","Type":"ContainerStarted","Data":"51176d3bd0cb309dad6e16d3458d8bcc3051d7df4a3df946f6c246e8a579cdbc"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.966530 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pcc6z" event={"ID":"b9b87214-6f46-4d58-b2fe-5618cfe4e0f0","Type":"ContainerStarted","Data":"0d85795e93c63f0816e55200bd2191db67947910e473d03b1bdf1167eb1dc345"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.966549 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pcc6z" event={"ID":"b9b87214-6f46-4d58-b2fe-5618cfe4e0f0","Type":"ContainerStarted","Data":"27a6017f2fc321b21d2b3c7101b553d8af1a06a2908e1fdee558e916c1651522"} Feb 14 11:12:30 crc kubenswrapper[4904]: I0214 11:12:30.986163 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bfn42"] Feb 14 11:12:31 crc kubenswrapper[4904]: I0214 11:12:31.008324 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qjrkr" Feb 14 11:12:31 crc kubenswrapper[4904]: I0214 11:12:31.052473 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:31 crc kubenswrapper[4904]: E0214 11:12:31.052783 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:31.552772528 +0000 UTC m=+142.365537189 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:31 crc kubenswrapper[4904]: I0214 11:12:31.056147 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:31 crc kubenswrapper[4904]: W0214 11:12:31.067980 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86a63698_caa8_4173_8f86_1bb56087babb.slice/crio-184e21a2374fbb0f16173c6266e96ad9d70599cb2e17f8453f6f2cb234c26925 WatchSource:0}: Error finding container 184e21a2374fbb0f16173c6266e96ad9d70599cb2e17f8453f6f2cb234c26925: Status 404 returned error can't find the container with id 184e21a2374fbb0f16173c6266e96ad9d70599cb2e17f8453f6f2cb234c26925 Feb 14 11:12:31 crc kubenswrapper[4904]: I0214 11:12:31.158231 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:31 crc kubenswrapper[4904]: E0214 11:12:31.158816 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:31.658800429 +0000 UTC m=+142.471565080 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:31 crc kubenswrapper[4904]: I0214 11:12:31.170867 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-qhldd" Feb 14 11:12:31 crc kubenswrapper[4904]: I0214 11:12:31.265826 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:31 crc kubenswrapper[4904]: E0214 11:12:31.266742 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:31.766725619 +0000 UTC m=+142.579490280 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:31 crc kubenswrapper[4904]: I0214 11:12:31.277794 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl"] Feb 14 11:12:31 crc kubenswrapper[4904]: W0214 11:12:31.295499 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc218d2bc_6512_4c03_b39e_05c801dac4b1.slice/crio-fe84cb5060e88639064b7432e38553f5d7de1a61afa26c811916bf3ba07f9b72 WatchSource:0}: Error finding container fe84cb5060e88639064b7432e38553f5d7de1a61afa26c811916bf3ba07f9b72: Status 404 returned error can't find the container with id fe84cb5060e88639064b7432e38553f5d7de1a61afa26c811916bf3ba07f9b72 Feb 14 11:12:31 crc kubenswrapper[4904]: I0214 11:12:31.367558 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:31 crc kubenswrapper[4904]: E0214 11:12:31.367859 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:31.867829929 +0000 UTC m=+142.680594590 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:31 crc kubenswrapper[4904]: I0214 11:12:31.478150 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-hgggh" podStartSLOduration=120.478116121 podStartE2EDuration="2m0.478116121s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:31.476448017 +0000 UTC m=+142.289212678" watchObservedRunningTime="2026-02-14 11:12:31.478116121 +0000 UTC m=+142.290880782" Feb 14 11:12:31 crc kubenswrapper[4904]: I0214 11:12:31.479058 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:31 crc kubenswrapper[4904]: E0214 11:12:31.479437 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:31.979425236 +0000 UTC m=+142.792189897 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:31 crc kubenswrapper[4904]: I0214 11:12:31.580358 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:31 crc kubenswrapper[4904]: E0214 11:12:31.581268 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:32.081243275 +0000 UTC m=+142.894007936 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:31 crc kubenswrapper[4904]: I0214 11:12:31.686613 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:31 crc kubenswrapper[4904]: E0214 11:12:31.687310 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:32.187294375 +0000 UTC m=+143.000059036 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:31 crc kubenswrapper[4904]: I0214 11:12:31.793056 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:31 crc kubenswrapper[4904]: E0214 11:12:31.794300 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:32.294284691 +0000 UTC m=+143.107049352 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:31 crc kubenswrapper[4904]: I0214 11:12:31.794324 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-x4nlj"] Feb 14 11:12:31 crc kubenswrapper[4904]: I0214 11:12:31.895428 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:31 crc kubenswrapper[4904]: E0214 11:12:31.916444 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:32.416416636 +0000 UTC m=+143.229181297 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:31 crc kubenswrapper[4904]: I0214 11:12:31.938344 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.015313 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:32 crc kubenswrapper[4904]: E0214 11:12:32.015869 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:32.515853782 +0000 UTC m=+143.328618443 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.023392 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n"] Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.059920 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-6wnjb" event={"ID":"c218d2bc-6512-4c03-b39e-05c801dac4b1","Type":"ContainerStarted","Data":"fe84cb5060e88639064b7432e38553f5d7de1a61afa26c811916bf3ba07f9b72"} Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.121186 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:32 crc kubenswrapper[4904]: E0214 11:12:32.123105 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:32.623090774 +0000 UTC m=+143.435855435 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.206235 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" event={"ID":"e831e54b-3826-4415-9b43-ce02e5ed81bf","Type":"ContainerStarted","Data":"84274d7ba89098ae8f02f769ec2f990db221ec0183ffe85b4292f4c5e2e92e3e"} Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.222588 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:32 crc kubenswrapper[4904]: E0214 11:12:32.222881 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:32.722867929 +0000 UTC m=+143.535632580 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.248218 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fqtd8" event={"ID":"11f24e26-e608-47a1-b8b1-8ac0159c1c5a","Type":"ContainerStarted","Data":"99d85f68e1e7f0a629ac50ab8eb94dd0bf6e933df576d89c49ff8ac4e0ef2ad8"} Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.273908 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-49jn7" event={"ID":"86a63698-caa8-4173-8f86-1bb56087babb","Type":"ContainerStarted","Data":"184e21a2374fbb0f16173c6266e96ad9d70599cb2e17f8453f6f2cb234c26925"} Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.302104 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4" event={"ID":"8b89659e-58f4-4c11-9839-2be0bdcd6e5e","Type":"ContainerStarted","Data":"4b44bacc66d5f642e505f8ca2f3ed2792bf06f1beaff9397c28b923651bf6cf5"} Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.330112 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g25mq" event={"ID":"fdb5ad89-a263-4f1f-bbe5-42bd5b91ea25","Type":"ContainerStarted","Data":"263809fab3d1cbed8a3e315f071a7d6e530f3a815c930dfb6e37a76a50253d89"} Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.330964 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:32 crc kubenswrapper[4904]: E0214 11:12:32.332092 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:32.832079593 +0000 UTC m=+143.644844254 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.338610 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-d948p" event={"ID":"30e3e22f-c17f-4efa-b5cd-33406a09def1","Type":"ContainerStarted","Data":"563aeb97131b4ee07b578110ec17ae4635a9e2217770c3b7765303ef418c966b"} Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.354875 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-2gsdv"] Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.379715 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7t272" event={"ID":"967509da-8c27-49be-aa96-3a7129937ddd","Type":"ContainerStarted","Data":"060d85065b53196006826ce14d8a2386fffc78cf556ceac89783a693218bee14"} Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.399065 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" podStartSLOduration=122.399050752 podStartE2EDuration="2m2.399050752s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:32.392319684 +0000 UTC m=+143.205084355" watchObservedRunningTime="2026-02-14 11:12:32.399050752 +0000 UTC m=+143.211815403" Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.401087 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7h7pz"] Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.402715 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52"] Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.405673 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-f95s6" Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.412146 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-7c6cp"] Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.417912 4904 patch_prober.go:28] interesting pod/console-operator-58897d9998-f95s6 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/readyz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.417971 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-f95s6" podUID="7c2c0ca2-4182-4c24-8710-c3b632e768f9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/readyz\": dial tcp 10.217.0.35:8443: connect: connection refused" Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.422526 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" event={"ID":"345bb34c-a9b2-4e88-94ce-3187b8933057","Type":"ContainerStarted","Data":"1a6e030b6b68bc65cdb0a0d991e171896cadbc9b7f3759adce0b0e4978475990"} Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.437263 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:32 crc kubenswrapper[4904]: E0214 11:12:32.438365 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:32.9383488 +0000 UTC m=+143.751113461 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.443035 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx"] Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.480295 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl" event={"ID":"652a99d8-a666-4c63-8bc5-71953e836b01","Type":"ContainerStarted","Data":"f8f1da4208edc78827f4fbaaedc4cf9df36925ea7353986b9c56d3694c1ac981"} Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.516077 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w" event={"ID":"ddd4a23a-8856-480c-91d6-829f766120a9","Type":"ContainerStarted","Data":"273e7a1b6e33cd99990acca2a58959daf2c98ec0ab9778ef3fd2252180926c61"} Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.537698 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ql7rx" event={"ID":"2201ed6d-9862-43d5-98ed-949539d6d0ae","Type":"ContainerStarted","Data":"f8bbeff98f3bdf349917146728dfb87c7205287370d29df8061c1372b3d1b64f"} Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.539988 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:32 crc kubenswrapper[4904]: E0214 11:12:32.540376 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:33.040363854 +0000 UTC m=+143.853128515 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.596607 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-f95s6" podStartSLOduration=122.596591129 podStartE2EDuration="2m2.596591129s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:32.594883174 +0000 UTC m=+143.407647835" watchObservedRunningTime="2026-02-14 11:12:32.596591129 +0000 UTC m=+143.409355790" Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.629568 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5pmk4" podStartSLOduration=121.629549259 podStartE2EDuration="2m1.629549259s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:32.628571013 +0000 UTC m=+143.441335674" watchObservedRunningTime="2026-02-14 11:12:32.629549259 +0000 UTC m=+143.442313920" Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.645768 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:32 crc kubenswrapper[4904]: E0214 11:12:32.656501 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:33.156481251 +0000 UTC m=+143.969245912 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.659952 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qsppc" event={"ID":"b42c5242-55b6-43aa-a9bf-e368d446c3a5","Type":"ContainerStarted","Data":"b4607ead404eb320bb9b1ac52d8964530715ec3a506b99499b1e463c453771be"} Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.663526 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-d948p" podStartSLOduration=121.663505826 podStartE2EDuration="2m1.663505826s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:32.659055308 +0000 UTC m=+143.471819969" watchObservedRunningTime="2026-02-14 11:12:32.663505826 +0000 UTC m=+143.476270487" Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.683118 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z" event={"ID":"66d09685-dfb3-4344-8cb2-f4e820e05bac","Type":"ContainerStarted","Data":"60b9f37eb9602e679007c680c45e5149964d3cc2b5c1d838cfc748df134cb07f"} Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.733650 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" event={"ID":"e3253b15-b01e-4c6b-91df-578ddb39b234","Type":"ContainerStarted","Data":"2bfeff960901204038f5df3ca1840f8355be562afe6d7f9d05596db2463f00f1"} Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.757648 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.764962 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" event={"ID":"de09ed81-3849-460e-93f2-fb0f5d72a4b9","Type":"ContainerStarted","Data":"51f4c3490e12bb9e23650a14e54d1bcc237b45431242c48956240f4fe42c4c35"} Feb 14 11:12:32 crc kubenswrapper[4904]: E0214 11:12:32.765221 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:33.265195131 +0000 UTC m=+144.077959792 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.765295 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.857603 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bzps5" event={"ID":"3cb9bb68-546e-4e52-9a15-5aeee129a5b4","Type":"ContainerStarted","Data":"85789abbd83491630b8e4478387297120ab3d1135f9912a7687e0842c125c053"} Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.859937 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-bzps5" Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.859966 4904 patch_prober.go:28] interesting pod/downloads-7954f5f757-bzps5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.859997 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bzps5" podUID="3cb9bb68-546e-4e52-9a15-5aeee129a5b4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.860260 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:32 crc kubenswrapper[4904]: E0214 11:12:32.860899 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:33.360882088 +0000 UTC m=+144.173646749 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.881930 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.964079 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjcqb"] Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.964689 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:32 crc kubenswrapper[4904]: E0214 11:12:32.965149 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:33.465133951 +0000 UTC m=+144.277898612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:32 crc kubenswrapper[4904]: I0214 11:12:32.966342 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9pzkg"] Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.060042 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c5fmb"] Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.062358 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ql7rx" podStartSLOduration=122.062339148 podStartE2EDuration="2m2.062339148s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:33.059119843 +0000 UTC m=+143.871884504" watchObservedRunningTime="2026-02-14 11:12:33.062339148 +0000 UTC m=+143.875103809" Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.066924 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:33 crc kubenswrapper[4904]: E0214 11:12:33.067291 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:33.567275109 +0000 UTC m=+144.380039770 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.110449 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-jq824" podStartSLOduration=123.110431818 podStartE2EDuration="2m3.110431818s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:33.108169328 +0000 UTC m=+143.920934009" watchObservedRunningTime="2026-02-14 11:12:33.110431818 +0000 UTC m=+143.923196479" Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.174151 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:33 crc kubenswrapper[4904]: E0214 11:12:33.174459 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:33.674449609 +0000 UTC m=+144.487214270 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.245003 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4n756"] Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.246655 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-bzps5" podStartSLOduration=123.246638935 podStartE2EDuration="2m3.246638935s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:33.235097671 +0000 UTC m=+144.047862332" watchObservedRunningTime="2026-02-14 11:12:33.246638935 +0000 UTC m=+144.059403596" Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.278018 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:33 crc kubenswrapper[4904]: E0214 11:12:33.278331 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:33.778314872 +0000 UTC m=+144.591079533 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.282257 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qsppc" podStartSLOduration=122.282239916 podStartE2EDuration="2m2.282239916s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:33.28167965 +0000 UTC m=+144.094444311" watchObservedRunningTime="2026-02-14 11:12:33.282239916 +0000 UTC m=+144.095004577" Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.291711 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-fjbmw"] Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.336264 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" podStartSLOduration=122.336248892 podStartE2EDuration="2m2.336248892s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:33.334754902 +0000 UTC m=+144.147519573" watchObservedRunningTime="2026-02-14 11:12:33.336248892 +0000 UTC m=+144.149013553" Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.380574 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:33 crc kubenswrapper[4904]: E0214 11:12:33.383005 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:33.882989766 +0000 UTC m=+144.695754427 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.449194 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-qjrkr"] Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.483220 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:33 crc kubenswrapper[4904]: E0214 11:12:33.483594 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:33.983577822 +0000 UTC m=+144.796342483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.531854 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-qhldd"] Feb 14 11:12:33 crc kubenswrapper[4904]: W0214 11:12:33.540504 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac1a67ad_be79_4f5f_adca_d2bb2d1bfac2.slice/crio-4c2440b51ee9235492efb239fc99d90f39c03ddda3715a9c48c0bfbb65813d0b WatchSource:0}: Error finding container 4c2440b51ee9235492efb239fc99d90f39c03ddda3715a9c48c0bfbb65813d0b: Status 404 returned error can't find the container with id 4c2440b51ee9235492efb239fc99d90f39c03ddda3715a9c48c0bfbb65813d0b Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.584396 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:33 crc kubenswrapper[4904]: E0214 11:12:33.584724 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:34.084711614 +0000 UTC m=+144.897476275 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.686470 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:33 crc kubenswrapper[4904]: E0214 11:12:33.687552 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:34.187531738 +0000 UTC m=+145.000296399 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.713234 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hfw59"] Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.769893 4904 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-5vpqp container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.769947 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" podUID="de09ed81-3849-460e-93f2-fb0f5d72a4b9" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.11:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.788728 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:33 crc kubenswrapper[4904]: E0214 11:12:33.789386 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:34.289373558 +0000 UTC m=+145.102138209 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:33 crc kubenswrapper[4904]: W0214 11:12:33.833673 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbbd4ae38_0244_445f_a4e8_6bffc1714d2d.slice/crio-ea48ad54d6198707b57460f9e6d54b9953aacc86102897c75d06838e4930582c WatchSource:0}: Error finding container ea48ad54d6198707b57460f9e6d54b9953aacc86102897c75d06838e4930582c: Status 404 returned error can't find the container with id ea48ad54d6198707b57460f9e6d54b9953aacc86102897c75d06838e4930582c Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.883216 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-f95s6" event={"ID":"7c2c0ca2-4182-4c24-8710-c3b632e768f9","Type":"ContainerStarted","Data":"c3c28864c9fe2b25a4d4fc50607ae64b697a26afafa6ca67496774b158294380"} Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.888233 4904 patch_prober.go:28] interesting pod/console-operator-58897d9998-f95s6 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/readyz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.888271 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-f95s6" podUID="7c2c0ca2-4182-4c24-8710-c3b632e768f9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/readyz\": dial tcp 10.217.0.35:8443: connect: connection refused" Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.889636 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:33 crc kubenswrapper[4904]: E0214 11:12:33.889978 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:34.389955035 +0000 UTC m=+145.202719696 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.898800 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl" event={"ID":"652a99d8-a666-4c63-8bc5-71953e836b01","Type":"ContainerStarted","Data":"2f7856266f24430cd76dba5646557b40345038f1e53d178d1c7c6a445333a04c"} Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.901155 4904 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-tbxbl container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.901202 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl" podUID="652a99d8-a666-4c63-8bc5-71953e836b01" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.901254 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl" Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.912822 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2q49s"] Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.930093 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx" event={"ID":"6d9f5d3d-fb18-4302-8b6b-62d4b8999b65","Type":"ContainerStarted","Data":"cb42d973f41b3aa1e0084c56c98b85cfe2206bc43df0751a876efe821a42fbc2"} Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.950015 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-x4nlj" event={"ID":"b1710cb1-bd05-4970-9b17-95cf75b9aec4","Type":"ContainerStarted","Data":"0780c68225da65052dfed64d4fb1e29b90e47daf8988a37d2af41f2dc0528b87"} Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.950056 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-x4nlj" event={"ID":"b1710cb1-bd05-4970-9b17-95cf75b9aec4","Type":"ContainerStarted","Data":"f5177cc32a909e611a742b6734ab781ff4878c8807e222826e9751b56e3b65b6"} Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.964902 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" event={"ID":"80b9d666-fa50-4d49-8c3f-efbd9d1247a9","Type":"ContainerStarted","Data":"d5fe28ecb4d43a1210487173eda4ac7f71711d7171acd5e2932ada1784d1c2ec"} Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.991461 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:33 crc kubenswrapper[4904]: E0214 11:12:33.992250 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:34.492230536 +0000 UTC m=+145.304995197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:33 crc kubenswrapper[4904]: I0214 11:12:33.995817 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z" event={"ID":"66d09685-dfb3-4344-8cb2-f4e820e05bac","Type":"ContainerStarted","Data":"9d8900372ae56af68d368aa30b0baae5d78d1a2223134a4a3ee030e49d06e498"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.013881 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qjrkr" event={"ID":"fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d","Type":"ContainerStarted","Data":"e74526ca2de5fd952d8d028e16d4bf18deb3c38c6be96115e5abf6bd9a06f7a7"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.031008 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" event={"ID":"e831e54b-3826-4415-9b43-ce02e5ed81bf","Type":"ContainerStarted","Data":"0a42a22bc2927f70424cd130fee97a557db035a17cdc29afff69644a2db61987"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.032220 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.042311 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hfw59" event={"ID":"bbd4ae38-0244-445f-a4e8-6bffc1714d2d","Type":"ContainerStarted","Data":"ea48ad54d6198707b57460f9e6d54b9953aacc86102897c75d06838e4930582c"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.042833 4904 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-bfn42 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.37:6443/healthz\": dial tcp 10.217.0.37:6443: connect: connection refused" start-of-body= Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.042882 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" podUID="e831e54b-3826-4415-9b43-ce02e5ed81bf" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.37:6443/healthz\": dial tcp 10.217.0.37:6443: connect: connection refused" Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.052467 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g25mq" event={"ID":"fdb5ad89-a263-4f1f-bbe5-42bd5b91ea25","Type":"ContainerStarted","Data":"760a6a7ea3c74018a12d2e888d1fdcf97ecf96d586a6b990397a8cf90941a8bb"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.070953 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl" podStartSLOduration=123.070934884 podStartE2EDuration="2m3.070934884s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:33.964152704 +0000 UTC m=+144.776917365" watchObservedRunningTime="2026-02-14 11:12:34.070934884 +0000 UTC m=+144.883699545" Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.073361 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-fjbmw" event={"ID":"ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2","Type":"ContainerStarted","Data":"4c2440b51ee9235492efb239fc99d90f39c03ddda3715a9c48c0bfbb65813d0b"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.074257 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7h7pz" event={"ID":"3b968f92-c773-4c84-bc7b-d00bf7a6f72b","Type":"ContainerStarted","Data":"1f7643ea906bd4c04539e8357ea1ccf0a85303bb9ab36a52c72328cb754a7191"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.082259 4904 generic.go:334] "Generic (PLEG): container finished" podID="345bb34c-a9b2-4e88-94ce-3187b8933057" containerID="1a6e030b6b68bc65cdb0a0d991e171896cadbc9b7f3759adce0b0e4978475990" exitCode=0 Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.082318 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" event={"ID":"345bb34c-a9b2-4e88-94ce-3187b8933057","Type":"ContainerDied","Data":"1a6e030b6b68bc65cdb0a0d991e171896cadbc9b7f3759adce0b0e4978475990"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.088480 4904 generic.go:334] "Generic (PLEG): container finished" podID="821984e5-b740-4332-9c06-1ccefda9a9fb" containerID="70a854fa18b833edbc0734bad46eef6480d87410a0a20ee1ac8f216a433513ee" exitCode=0 Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.088565 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" event={"ID":"821984e5-b740-4332-9c06-1ccefda9a9fb","Type":"ContainerDied","Data":"70a854fa18b833edbc0734bad46eef6480d87410a0a20ee1ac8f216a433513ee"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.093284 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7c6cp" event={"ID":"261ad318-57b3-48ca-8dac-e8bd53528694","Type":"ContainerStarted","Data":"7fccd0ed3d7eaaf404719a88d37ef0cf98a037d712dbc05f1db3e966cd5bcbfd"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.093890 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:34 crc kubenswrapper[4904]: E0214 11:12:34.094908 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:34.594893076 +0000 UTC m=+145.407657737 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.109950 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pcc6z" event={"ID":"b9b87214-6f46-4d58-b2fe-5618cfe4e0f0","Type":"ContainerStarted","Data":"29ddd171a318113d366c769c909d1597e5d0b6646617c870b09ae8d945fa5e99"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.117949 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g25mq" podStartSLOduration=124.117931134 podStartE2EDuration="2m4.117931134s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:34.117342599 +0000 UTC m=+144.930107260" watchObservedRunningTime="2026-02-14 11:12:34.117931134 +0000 UTC m=+144.930695795" Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.118596 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" podStartSLOduration=124.118592702 podStartE2EDuration="2m4.118592702s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:34.069744432 +0000 UTC m=+144.882509093" watchObservedRunningTime="2026-02-14 11:12:34.118592702 +0000 UTC m=+144.931357363" Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.123546 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-qh6sr" event={"ID":"91b7cb2e-1781-41da-9c90-13b40be0a2b4","Type":"ContainerStarted","Data":"e99eddd25649647f12b41d8b0808ad14c2d1bfd1a173a849cce6d228e3132cec"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.127631 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-6wnjb" event={"ID":"c218d2bc-6512-4c03-b39e-05c801dac4b1","Type":"ContainerStarted","Data":"df34bd2aa899287b2fb965b5f407cfe414bc4e46081c5f8fbd6c26264647690d"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.146821 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4n756" event={"ID":"e9d77fde-3517-4aea-abcb-055541763e48","Type":"ContainerStarted","Data":"2a519bedf4609fc5b78cafbff787733bdf6ec51c92114040393f1c6fb9331fba"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.195462 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:34 crc kubenswrapper[4904]: E0214 11:12:34.197352 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:34.697341372 +0000 UTC m=+145.510106033 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.242244 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" event={"ID":"d910fb92-a40a-4167-be6f-113f22c206b2","Type":"ContainerStarted","Data":"02efeac069d631dbfd5fb9698eb44b649260f643dc8eadeba4b881ddbcb02c75"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.243083 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.243997 4904 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-rnf52 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.244036 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" podUID="d910fb92-a40a-4167-be6f-113f22c206b2" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.256516 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pcc6z" podStartSLOduration=124.256496414 podStartE2EDuration="2m4.256496414s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:34.192888914 +0000 UTC m=+145.005653575" watchObservedRunningTime="2026-02-14 11:12:34.256496414 +0000 UTC m=+145.069261075" Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.264231 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" event={"ID":"f130a8a5-24b3-4916-a0c7-b1fa5e55cad4","Type":"ContainerStarted","Data":"d3b7f6adb60da3cbff5abdf8689b57ff9d0e31b8d04cbcb123d186c35c73c7a4"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.287278 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qhldd" event={"ID":"966787c0-c9ad-460e-be70-26cb6acbcf0f","Type":"ContainerStarted","Data":"595ecf6afac51bd2ddf9abe12c265804ae4f869012fd2be6b33ab50b0b28c2c2"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.299129 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:34 crc kubenswrapper[4904]: E0214 11:12:34.299305 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:34.799291084 +0000 UTC m=+145.612055745 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.300002 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:34 crc kubenswrapper[4904]: E0214 11:12:34.300374 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:34.800359463 +0000 UTC m=+145.613124124 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.362232 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n" event={"ID":"5db3acfa-ac25-413e-ac37-4a782745ab08","Type":"ContainerStarted","Data":"c175dbd4847594b5b5fc5baa40ce4462ca41c947cd84b0f25d16df2d61fa085d"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.363263 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n" Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.364591 4904 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-zpj2n container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.364639 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n" podUID="5db3acfa-ac25-413e-ac37-4a782745ab08" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.377758 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" podStartSLOduration=123.377743896 podStartE2EDuration="2m3.377743896s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:34.368158183 +0000 UTC m=+145.180922844" watchObservedRunningTime="2026-02-14 11:12:34.377743896 +0000 UTC m=+145.190508557" Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.402446 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:34 crc kubenswrapper[4904]: E0214 11:12:34.402735 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:34.902722106 +0000 UTC m=+145.715486767 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.403737 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjcqb" event={"ID":"ca75271d-7ae7-4a71-af10-732866e33d74","Type":"ContainerStarted","Data":"0cfecbd435efefef321bd7ba7a4bcd69da760b8d1ef21a0dce02e5ca2cba7084"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.434486 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-6wnjb" podStartSLOduration=7.434467744 podStartE2EDuration="7.434467744s" podCreationTimestamp="2026-02-14 11:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:34.433117328 +0000 UTC m=+145.245881979" watchObservedRunningTime="2026-02-14 11:12:34.434467744 +0000 UTC m=+145.247232405" Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.439016 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2gsdv" event={"ID":"5897936b-fc22-4350-a0f8-c4397854da88","Type":"ContainerStarted","Data":"12023f67e42621f523c0c2fac6456da2068f076f2c24415077d928e9fb4f3dc9"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.520283 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n" podStartSLOduration=123.52025703 podStartE2EDuration="2m3.52025703s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:34.499424 +0000 UTC m=+145.312188661" watchObservedRunningTime="2026-02-14 11:12:34.52025703 +0000 UTC m=+145.333021691" Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.536378 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:34 crc kubenswrapper[4904]: E0214 11:12:34.536723 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:35.036709404 +0000 UTC m=+145.849474065 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.639616 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:34 crc kubenswrapper[4904]: E0214 11:12:34.639922 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:35.139907199 +0000 UTC m=+145.952671850 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.686849 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fqtd8" event={"ID":"11f24e26-e608-47a1-b8b1-8ac0159c1c5a","Type":"ContainerStarted","Data":"801dc1a99bedf3e80801980ff035ef4f1765a1fd56d24cb4c006a0341a377f62"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.742099 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:34 crc kubenswrapper[4904]: E0214 11:12:34.742436 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:35.242414846 +0000 UTC m=+146.055179507 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.805019 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" event={"ID":"a7f9f078-ff94-4adc-be45-927915d6e55f","Type":"ContainerStarted","Data":"d781427f42601807956ef314152870ee7733d6d19b9f5f163971df1ae632d117"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.832812 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c5fmb" event={"ID":"477826a2-46cd-4145-afc9-92b69930526e","Type":"ContainerStarted","Data":"90d84b3fa36c1c1779c7bc01f33e97ba5e6e816d19cf1dd18cf35082bf405a3c"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.835220 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-49jn7" event={"ID":"86a63698-caa8-4173-8f86-1bb56087babb","Type":"ContainerStarted","Data":"f3b2405f7b2d981c46c2e19cbbc248214c81d67bfb2ac32c55832b55b2ce4655"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.848515 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w" event={"ID":"ddd4a23a-8856-480c-91d6-829f766120a9","Type":"ContainerStarted","Data":"5aceea35dea3ae7a66d11c0f4bc58cd018e648498d8711606b3f49bc0fc4ad13"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.850433 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:34 crc kubenswrapper[4904]: E0214 11:12:34.854587 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:35.354565408 +0000 UTC m=+146.167330069 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.910148 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-d9d4k" podStartSLOduration=124.910121265 podStartE2EDuration="2m4.910121265s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:34.907516006 +0000 UTC m=+145.720280657" watchObservedRunningTime="2026-02-14 11:12:34.910121265 +0000 UTC m=+145.722885926" Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.921626 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7t272" event={"ID":"967509da-8c27-49be-aa96-3a7129937ddd","Type":"ContainerStarted","Data":"c3c9b55fbfea53a51092396b348b88932f7ae7ec0099b710ced917d2a274e965"} Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.924667 4904 patch_prober.go:28] interesting pod/downloads-7954f5f757-bzps5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.924697 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bzps5" podUID="3cb9bb68-546e-4e52-9a15-5aeee129a5b4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.954044 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:34 crc kubenswrapper[4904]: E0214 11:12:34.956087 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:35.456076559 +0000 UTC m=+146.268841220 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:34 crc kubenswrapper[4904]: I0214 11:12:34.993891 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5vpqp" Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.055409 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:35 crc kubenswrapper[4904]: E0214 11:12:35.056984 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:35.556964013 +0000 UTC m=+146.369728674 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.078224 4904 csr.go:261] certificate signing request csr-nb4p7 is approved, waiting to be issued Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.106473 4904 csr.go:257] certificate signing request csr-nb4p7 is issued Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.125455 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vdl4w" podStartSLOduration=125.125439592 podStartE2EDuration="2m5.125439592s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:35.011904853 +0000 UTC m=+145.824669514" watchObservedRunningTime="2026-02-14 11:12:35.125439592 +0000 UTC m=+145.938204253" Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.159394 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:35 crc kubenswrapper[4904]: E0214 11:12:35.159698 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:35.659687486 +0000 UTC m=+146.472452147 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.199862 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-49jn7" podStartSLOduration=125.199827306 podStartE2EDuration="2m5.199827306s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:35.133811303 +0000 UTC m=+145.946575964" watchObservedRunningTime="2026-02-14 11:12:35.199827306 +0000 UTC m=+146.012591967" Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.264340 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:35 crc kubenswrapper[4904]: E0214 11:12:35.264722 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:35.764700619 +0000 UTC m=+146.577465280 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.366884 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:35 crc kubenswrapper[4904]: E0214 11:12:35.367593 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:35.867581726 +0000 UTC m=+146.680346387 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.469800 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:35 crc kubenswrapper[4904]: E0214 11:12:35.470159 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:35.970143485 +0000 UTC m=+146.782908146 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.571538 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:35 crc kubenswrapper[4904]: E0214 11:12:35.572026 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:36.072011005 +0000 UTC m=+146.884775676 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.672927 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:35 crc kubenswrapper[4904]: E0214 11:12:35.673362 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:36.173348781 +0000 UTC m=+146.986113442 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.673655 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:35 crc kubenswrapper[4904]: E0214 11:12:35.673989 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:36.173981478 +0000 UTC m=+146.986746139 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.774659 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:35 crc kubenswrapper[4904]: E0214 11:12:35.775037 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:36.275019826 +0000 UTC m=+147.087784487 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.855180 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7t272" podStartSLOduration=125.855164702 podStartE2EDuration="2m5.855164702s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:35.273777429 +0000 UTC m=+146.086542100" watchObservedRunningTime="2026-02-14 11:12:35.855164702 +0000 UTC m=+146.667929363" Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.858422 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.858962 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.865079 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.865667 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.882031 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:35 crc kubenswrapper[4904]: E0214 11:12:35.882431 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:36.382413643 +0000 UTC m=+147.195178304 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.912312 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.933502 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4n756" event={"ID":"e9d77fde-3517-4aea-abcb-055541763e48","Type":"ContainerStarted","Data":"ffe4186f5fa78512f3a5939319556aa9014a394364dd250a6153bd1eaa71021b"} Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.959600 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" event={"ID":"d910fb92-a40a-4167-be6f-113f22c206b2","Type":"ContainerStarted","Data":"976a29dc369aa3b647f0f609fcfc3b5caccfdfc11a27fda35184577280c3470b"} Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.986018 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.986252 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6e37b832-8ee6-4eec-b048-3517ae74ed77-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6e37b832-8ee6-4eec-b048-3517ae74ed77\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.986310 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e37b832-8ee6-4eec-b048-3517ae74ed77-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6e37b832-8ee6-4eec-b048-3517ae74ed77\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 14 11:12:35 crc kubenswrapper[4904]: E0214 11:12:35.986439 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:36.486423779 +0000 UTC m=+147.299188440 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.991906 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c5fmb" event={"ID":"477826a2-46cd-4145-afc9-92b69930526e","Type":"ContainerStarted","Data":"6338fbf7ff6257162fdabd4885bc80fb652ef7f30eafea3374d098d57c3375eb"} Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.997297 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2q49s" event={"ID":"b6796678-081b-4588-88d2-1cf0ef013ba4","Type":"ContainerStarted","Data":"554a70f115f2fc27955562120db137922818c7834e7385ff1da8d7af0f202fb6"} Feb 14 11:12:35 crc kubenswrapper[4904]: I0214 11:12:35.997335 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2q49s" event={"ID":"b6796678-081b-4588-88d2-1cf0ef013ba4","Type":"ContainerStarted","Data":"53cf59c74af9a9ea7bfcfc26dc9830dc953aea030d0ecf8ffd19cadd92e6d37e"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.001607 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z" event={"ID":"66d09685-dfb3-4344-8cb2-f4e820e05bac","Type":"ContainerStarted","Data":"4ef2ad6997bca47529d4b70428acfcf02f38c4dce223e5b3127229be2b96fb22"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.007577 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-fjbmw" event={"ID":"ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2","Type":"ContainerStarted","Data":"5117e8a5ef4f0d0ab3cd472750f4cab823e0610bd37ea7c6f4a4d77314c682fe"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.022475 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qjrkr" event={"ID":"fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d","Type":"ContainerStarted","Data":"5b192c569f54d622b945534d3d8075fa7115c317258f4e74f7057b0baa80cb54"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.022514 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qjrkr" event={"ID":"fb6638cc-49f1-43d6-ab9e-4f0c2a737b7d","Type":"ContainerStarted","Data":"6f36465af68754d33d470fe96c12e48abf9e66e8acf619be83b7ee9d751d8f40"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.041371 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7h7pz" event={"ID":"3b968f92-c773-4c84-bc7b-d00bf7a6f72b","Type":"ContainerStarted","Data":"dff24dcfd0f0a95b202c5adc9773a062d4ee9aeed33ec75098bd59b1f2e8573e"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.041416 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7h7pz" event={"ID":"3b968f92-c773-4c84-bc7b-d00bf7a6f72b","Type":"ContainerStarted","Data":"01a9ccf4713374bf1fccbb31fdad193ce2c5519e1f0dc656cafefcc993a55271"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.048906 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" event={"ID":"f130a8a5-24b3-4916-a0c7-b1fa5e55cad4","Type":"ContainerStarted","Data":"c58cdbc463ae1de3ed0c35e320580faeaddb6a95577c99c69fe14ae19e9381c9"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.049711 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.051494 4904 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9pzkg container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.051536 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" podUID="f130a8a5-24b3-4916-a0c7-b1fa5e55cad4" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.053977 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4n756" podStartSLOduration=125.053966313 podStartE2EDuration="2m5.053966313s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:36.028067539 +0000 UTC m=+146.840832200" watchObservedRunningTime="2026-02-14 11:12:36.053966313 +0000 UTC m=+146.866730974" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.060077 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-qh6sr" event={"ID":"91b7cb2e-1781-41da-9c90-13b40be0a2b4","Type":"ContainerStarted","Data":"ea6ebb302088aa4fc65077b4227e5cb86c99ca6ac262dd188eca2aaae1f89d4c"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.067207 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n" event={"ID":"5db3acfa-ac25-413e-ac37-4a782745ab08","Type":"ContainerStarted","Data":"282f10dd6f91bbaa39eb756a95f540645779a6e0db66ac8e59edd6bbaae39d0b"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.070747 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2gsdv" event={"ID":"5897936b-fc22-4350-a0f8-c4397854da88","Type":"ContainerStarted","Data":"408bec2db02422c860a916f6a4bde743903c63785e395f1aea0a53fcf0d054c2"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.070786 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2gsdv" event={"ID":"5897936b-fc22-4350-a0f8-c4397854da88","Type":"ContainerStarted","Data":"d0931e78dc930bc17fff89908917bfc537d71903676f42bcd523dc0cb8744548"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.088820 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6e37b832-8ee6-4eec-b048-3517ae74ed77-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6e37b832-8ee6-4eec-b048-3517ae74ed77\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.088924 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e37b832-8ee6-4eec-b048-3517ae74ed77-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6e37b832-8ee6-4eec-b048-3517ae74ed77\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.089037 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:36 crc kubenswrapper[4904]: E0214 11:12:36.089309 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:36.589297005 +0000 UTC m=+147.402061656 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.090394 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" event={"ID":"80b9d666-fa50-4d49-8c3f-efbd9d1247a9","Type":"ContainerStarted","Data":"2712310893b674d60bda85022d77761e67bbea9eaad73b9db2d5bbc38ea56969"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.094377 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6e37b832-8ee6-4eec-b048-3517ae74ed77-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6e37b832-8ee6-4eec-b048-3517ae74ed77\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.126289 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zpj2n" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.127150 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-14 11:07:35 +0000 UTC, rotation deadline is 2026-11-20 07:45:08.175120865 +0000 UTC Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.127179 4904 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6692h32m32.047944109s for next certificate rotation Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.147095 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" event={"ID":"345bb34c-a9b2-4e88-94ce-3187b8933057","Type":"ContainerStarted","Data":"882b5018fd230ee712bfe84315dc152109b09cba0b848cfd5cc64d524ddae0a5"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.147707 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.151525 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e37b832-8ee6-4eec-b048-3517ae74ed77-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6e37b832-8ee6-4eec-b048-3517ae74ed77\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.156311 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6rh7z" podStartSLOduration=125.156293825 podStartE2EDuration="2m5.156293825s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:36.05461963 +0000 UTC m=+146.867384291" watchObservedRunningTime="2026-02-14 11:12:36.156293825 +0000 UTC m=+146.969058486" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.161647 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fqtd8" event={"ID":"11f24e26-e608-47a1-b8b1-8ac0159c1c5a","Type":"ContainerStarted","Data":"2c8fc3410d410e83e06bd4f548bb956de2067c2179dccf6b2e04b558e0e7027e"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.162502 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fqtd8" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.178441 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" event={"ID":"821984e5-b740-4332-9c06-1ccefda9a9fb","Type":"ContainerStarted","Data":"64673ffd651c3395783607417eace27a7dfd93ef82331b299bf328004b358746"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.185596 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.188125 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hfw59" event={"ID":"bbd4ae38-0244-445f-a4e8-6bffc1714d2d","Type":"ContainerStarted","Data":"80eeb93d04fef7e2be675f9520d08fe07d420d60e1153be3ec3070a92358bf3f"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.192274 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:36 crc kubenswrapper[4904]: E0214 11:12:36.192405 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:36.692382008 +0000 UTC m=+147.505146669 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.192700 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:36 crc kubenswrapper[4904]: E0214 11:12:36.195196 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:36.695180882 +0000 UTC m=+147.507945543 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.214395 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-x4nlj" event={"ID":"b1710cb1-bd05-4970-9b17-95cf75b9aec4","Type":"ContainerStarted","Data":"836027e72bdbb50dddeef301e9b7aa0a74a6cf1c23ae6b55e824869c191f0901"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.228059 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx" event={"ID":"6d9f5d3d-fb18-4302-8b6b-62d4b8999b65","Type":"ContainerStarted","Data":"6a6ed2121e9a25309d4d0411dd1abe47ea2f42a715eae146c16d359ff323786c"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.245088 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7c6cp" event={"ID":"261ad318-57b3-48ca-8dac-e8bd53528694","Type":"ContainerStarted","Data":"0ceb5bec8f16f4e10c8554e8100955c5da044ec2434441963b8f51cf7c0d110a"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.257413 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjcqb" event={"ID":"ca75271d-7ae7-4a71-af10-732866e33d74","Type":"ContainerStarted","Data":"88ccdb25a81699ca0f40c5a741b964e6a92b03247a8041a2c78cdcead62e4374"} Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.260663 4904 patch_prober.go:28] interesting pod/downloads-7954f5f757-bzps5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.260796 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bzps5" podUID="3cb9bb68-546e-4e52-9a15-5aeee129a5b4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.262736 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qjrkr" podStartSLOduration=125.262727125 podStartE2EDuration="2m5.262727125s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:36.158871273 +0000 UTC m=+146.971635934" watchObservedRunningTime="2026-02-14 11:12:36.262727125 +0000 UTC m=+147.075491786" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.266079 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c5fmb" podStartSLOduration=125.266072534 podStartE2EDuration="2m5.266072534s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:36.250957785 +0000 UTC m=+147.063722446" watchObservedRunningTime="2026-02-14 11:12:36.266072534 +0000 UTC m=+147.078837195" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.283478 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tbxbl" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.295821 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:36 crc kubenswrapper[4904]: E0214 11:12:36.297258 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:36.797241687 +0000 UTC m=+147.610006348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.401021 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:36 crc kubenswrapper[4904]: E0214 11:12:36.405032 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:36.905020694 +0000 UTC m=+147.717785355 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.506312 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sjcqb" podStartSLOduration=125.506298068 podStartE2EDuration="2m5.506298068s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:36.505443676 +0000 UTC m=+147.318208337" watchObservedRunningTime="2026-02-14 11:12:36.506298068 +0000 UTC m=+147.319062729" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.506456 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:36 crc kubenswrapper[4904]: E0214 11:12:36.506568 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.006534734 +0000 UTC m=+147.819299385 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.510382 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:36 crc kubenswrapper[4904]: E0214 11:12:36.511555 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.011451424 +0000 UTC m=+147.824216085 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.521164 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-2q49s" podStartSLOduration=9.52114104 podStartE2EDuration="9.52114104s" podCreationTimestamp="2026-02-14 11:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:36.399720053 +0000 UTC m=+147.212484714" watchObservedRunningTime="2026-02-14 11:12:36.52114104 +0000 UTC m=+147.333905701" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.612505 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:36 crc kubenswrapper[4904]: E0214 11:12:36.612847 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.112817991 +0000 UTC m=+147.925582652 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.714242 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:36 crc kubenswrapper[4904]: E0214 11:12:36.714690 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.214674391 +0000 UTC m=+148.027439052 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.749603 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx" podStartSLOduration=126.749574603 podStartE2EDuration="2m6.749574603s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:36.62945256 +0000 UTC m=+147.442217221" watchObservedRunningTime="2026-02-14 11:12:36.749574603 +0000 UTC m=+147.562339264" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.814883 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:36 crc kubenswrapper[4904]: E0214 11:12:36.815104 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.315062852 +0000 UTC m=+148.127827513 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.815264 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:36 crc kubenswrapper[4904]: E0214 11:12:36.815607 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.315594476 +0000 UTC m=+148.128359127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.821024 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" podStartSLOduration=126.821012939 podStartE2EDuration="2m6.821012939s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:36.758370305 +0000 UTC m=+147.571134986" watchObservedRunningTime="2026-02-14 11:12:36.821012939 +0000 UTC m=+147.633777600" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.822721 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" podStartSLOduration=126.822713554 podStartE2EDuration="2m6.822713554s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:36.80856368 +0000 UTC m=+147.621328341" watchObservedRunningTime="2026-02-14 11:12:36.822713554 +0000 UTC m=+147.635478215" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.824764 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.844182 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-f95s6" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.854348 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-qh6sr" podStartSLOduration=125.854332829 podStartE2EDuration="2m5.854332829s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:36.852911582 +0000 UTC m=+147.665676243" watchObservedRunningTime="2026-02-14 11:12:36.854332829 +0000 UTC m=+147.667097490" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.916695 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:36 crc kubenswrapper[4904]: E0214 11:12:36.917167 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.417150558 +0000 UTC m=+148.229915219 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.917280 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:36 crc kubenswrapper[4904]: E0214 11:12:36.917565 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.417559059 +0000 UTC m=+148.230323710 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.960860 4904 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-rnf52 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.960929 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" podUID="d910fb92-a40a-4167-be6f-113f22c206b2" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 14 11:12:36 crc kubenswrapper[4904]: I0214 11:12:36.993338 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" podStartSLOduration=125.99331971 podStartE2EDuration="2m5.99331971s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:36.993079103 +0000 UTC m=+147.805843764" watchObservedRunningTime="2026-02-14 11:12:36.99331971 +0000 UTC m=+147.806084371" Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.018008 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:37 crc kubenswrapper[4904]: E0214 11:12:37.018187 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.518159616 +0000 UTC m=+148.330924277 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.018303 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:37 crc kubenswrapper[4904]: E0214 11:12:37.018620 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.518608527 +0000 UTC m=+148.331373188 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.057112 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.062287 4904 patch_prober.go:28] interesting pod/router-default-5444994796-qh6sr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 14 11:12:37 crc kubenswrapper[4904]: [-]has-synced failed: reason withheld Feb 14 11:12:37 crc kubenswrapper[4904]: [+]process-running ok Feb 14 11:12:37 crc kubenswrapper[4904]: healthz check failed Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.062329 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qh6sr" podUID="91b7cb2e-1781-41da-9c90-13b40be0a2b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.119798 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:37 crc kubenswrapper[4904]: E0214 11:12:37.119973 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.619948613 +0000 UTC m=+148.432713274 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.120178 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:37 crc kubenswrapper[4904]: E0214 11:12:37.120492 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.620485068 +0000 UTC m=+148.433249729 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.189798 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-7h7pz" podStartSLOduration=127.189781097 podStartE2EDuration="2m7.189781097s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:37.188239467 +0000 UTC m=+148.001004138" watchObservedRunningTime="2026-02-14 11:12:37.189781097 +0000 UTC m=+148.002545758" Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.221526 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:37 crc kubenswrapper[4904]: E0214 11:12:37.221660 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.721642749 +0000 UTC m=+148.534407410 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.222041 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:37 crc kubenswrapper[4904]: E0214 11:12:37.222327 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.722316437 +0000 UTC m=+148.535081098 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.258845 4904 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-bfn42 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.37:6443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.258919 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" podUID="e831e54b-3826-4415-9b43-ce02e5ed81bf" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.37:6443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.286412 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qhldd" event={"ID":"966787c0-c9ad-460e-be70-26cb6acbcf0f","Type":"ContainerStarted","Data":"9fd6e0e47a24eea459d785ef181c532b9a25becb3a7d6694121bbf055a9de964"} Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.288301 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6e37b832-8ee6-4eec-b048-3517ae74ed77","Type":"ContainerStarted","Data":"ab71ba87990201eb4e7cd757ed766aa61a3b86945930103e35929b849e440296"} Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.306294 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-fjbmw" event={"ID":"ac1a67ad-be79-4f5f-adca-d2bb2d1bfac2","Type":"ContainerStarted","Data":"3853b3efb2c2355b57c82bbff548503dd203e6473647e2b3ca85a3478f9ff6c8"} Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.307904 4904 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9pzkg container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.308032 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" podUID="f130a8a5-24b3-4916-a0c7-b1fa5e55cad4" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.315800 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.323560 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:37 crc kubenswrapper[4904]: E0214 11:12:37.323879 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.823864149 +0000 UTC m=+148.636628810 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.424730 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:37 crc kubenswrapper[4904]: E0214 11:12:37.429521 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:37.929509519 +0000 UTC m=+148.742274180 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.526457 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:37 crc kubenswrapper[4904]: E0214 11:12:37.527078 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:38.027063235 +0000 UTC m=+148.839827886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.575112 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-x4nlj" podStartSLOduration=126.575094464 podStartE2EDuration="2m6.575094464s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:37.507008455 +0000 UTC m=+148.319773116" watchObservedRunningTime="2026-02-14 11:12:37.575094464 +0000 UTC m=+148.387859125" Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.576613 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2gsdv" podStartSLOduration=126.576607394 podStartE2EDuration="2m6.576607394s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:37.387963062 +0000 UTC m=+148.200727723" watchObservedRunningTime="2026-02-14 11:12:37.576607394 +0000 UTC m=+148.389372055" Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.627966 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:37 crc kubenswrapper[4904]: E0214 11:12:37.628358 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:38.12834254 +0000 UTC m=+148.941107201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.729338 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:37 crc kubenswrapper[4904]: E0214 11:12:37.729619 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:38.229603254 +0000 UTC m=+149.042367915 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.831223 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:37 crc kubenswrapper[4904]: E0214 11:12:37.831868 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:38.331855174 +0000 UTC m=+149.144619835 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.902349 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-7c6cp" podStartSLOduration=127.902332715 podStartE2EDuration="2m7.902332715s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:37.820274669 +0000 UTC m=+148.633039330" watchObservedRunningTime="2026-02-14 11:12:37.902332715 +0000 UTC m=+148.715097376" Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.904781 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bvpm5"] Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.905682 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bvpm5" Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.932187 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.932347 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:12:37 crc kubenswrapper[4904]: E0214 11:12:37.932416 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:38.432393119 +0000 UTC m=+149.245157780 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.932567 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.932629 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.932656 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:12:37 crc kubenswrapper[4904]: E0214 11:12:37.932905 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:38.432889342 +0000 UTC m=+149.245653993 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.933349 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.934592 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.938551 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.938674 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.956018 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.960252 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 14 11:12:37 crc kubenswrapper[4904]: I0214 11:12:37.969205 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.029166 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bvpm5"] Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.029388 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fqtd8" podStartSLOduration=127.02938049 podStartE2EDuration="2m7.02938049s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:37.987581706 +0000 UTC m=+148.800346367" watchObservedRunningTime="2026-02-14 11:12:38.02938049 +0000 UTC m=+148.842145151" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.035516 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.035694 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63c9ce1a-01e3-42f3-8027-bfda9be0b3ff-catalog-content\") pod \"community-operators-bvpm5\" (UID: \"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff\") " pod="openshift-marketplace/community-operators-bvpm5" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.035717 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63c9ce1a-01e3-42f3-8027-bfda9be0b3ff-utilities\") pod \"community-operators-bvpm5\" (UID: \"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff\") " pod="openshift-marketplace/community-operators-bvpm5" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.035745 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kh6f\" (UniqueName: \"kubernetes.io/projected/63c9ce1a-01e3-42f3-8027-bfda9be0b3ff-kube-api-access-6kh6f\") pod \"community-operators-bvpm5\" (UID: \"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff\") " pod="openshift-marketplace/community-operators-bvpm5" Feb 14 11:12:38 crc kubenswrapper[4904]: E0214 11:12:38.035871 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:38.535857412 +0000 UTC m=+149.348622073 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.072617 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-977kk"] Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.074077 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-977kk" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.073107 4904 patch_prober.go:28] interesting pod/router-default-5444994796-qh6sr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 14 11:12:38 crc kubenswrapper[4904]: [-]has-synced failed: reason withheld Feb 14 11:12:38 crc kubenswrapper[4904]: [+]process-running ok Feb 14 11:12:38 crc kubenswrapper[4904]: healthz check failed Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.076291 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qh6sr" podUID="91b7cb2e-1781-41da-9c90-13b40be0a2b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.086916 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.090381 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-977kk"] Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.136887 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63c9ce1a-01e3-42f3-8027-bfda9be0b3ff-utilities\") pod \"community-operators-bvpm5\" (UID: \"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff\") " pod="openshift-marketplace/community-operators-bvpm5" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.136927 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kh6f\" (UniqueName: \"kubernetes.io/projected/63c9ce1a-01e3-42f3-8027-bfda9be0b3ff-kube-api-access-6kh6f\") pod \"community-operators-bvpm5\" (UID: \"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff\") " pod="openshift-marketplace/community-operators-bvpm5" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.136976 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92dc9e32-c069-4de6-a3ae-28b80b60d722-utilities\") pod \"certified-operators-977kk\" (UID: \"92dc9e32-c069-4de6-a3ae-28b80b60d722\") " pod="openshift-marketplace/certified-operators-977kk" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.137025 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw586\" (UniqueName: \"kubernetes.io/projected/92dc9e32-c069-4de6-a3ae-28b80b60d722-kube-api-access-vw586\") pod \"certified-operators-977kk\" (UID: \"92dc9e32-c069-4de6-a3ae-28b80b60d722\") " pod="openshift-marketplace/certified-operators-977kk" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.137057 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.137074 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92dc9e32-c069-4de6-a3ae-28b80b60d722-catalog-content\") pod \"certified-operators-977kk\" (UID: \"92dc9e32-c069-4de6-a3ae-28b80b60d722\") " pod="openshift-marketplace/certified-operators-977kk" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.137094 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63c9ce1a-01e3-42f3-8027-bfda9be0b3ff-catalog-content\") pod \"community-operators-bvpm5\" (UID: \"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff\") " pod="openshift-marketplace/community-operators-bvpm5" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.137448 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63c9ce1a-01e3-42f3-8027-bfda9be0b3ff-catalog-content\") pod \"community-operators-bvpm5\" (UID: \"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff\") " pod="openshift-marketplace/community-operators-bvpm5" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.137660 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63c9ce1a-01e3-42f3-8027-bfda9be0b3ff-utilities\") pod \"community-operators-bvpm5\" (UID: \"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff\") " pod="openshift-marketplace/community-operators-bvpm5" Feb 14 11:12:38 crc kubenswrapper[4904]: E0214 11:12:38.138044 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:38.63803396 +0000 UTC m=+149.450798631 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.156465 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.197281 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zx9lt"] Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.198578 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx9lt" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.200981 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" podStartSLOduration=127.200971522 podStartE2EDuration="2m7.200971522s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:38.19941788 +0000 UTC m=+149.012182541" watchObservedRunningTime="2026-02-14 11:12:38.200971522 +0000 UTC m=+149.013736183" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.207539 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kh6f\" (UniqueName: \"kubernetes.io/projected/63c9ce1a-01e3-42f3-8027-bfda9be0b3ff-kube-api-access-6kh6f\") pod \"community-operators-bvpm5\" (UID: \"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff\") " pod="openshift-marketplace/community-operators-bvpm5" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.221822 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bvpm5" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.223029 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zx9lt"] Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.237498 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.237765 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee507b1a-9072-40b2-8450-c57209c10a39-utilities\") pod \"community-operators-zx9lt\" (UID: \"ee507b1a-9072-40b2-8450-c57209c10a39\") " pod="openshift-marketplace/community-operators-zx9lt" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.237799 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92dc9e32-c069-4de6-a3ae-28b80b60d722-catalog-content\") pod \"certified-operators-977kk\" (UID: \"92dc9e32-c069-4de6-a3ae-28b80b60d722\") " pod="openshift-marketplace/certified-operators-977kk" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.237864 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92dc9e32-c069-4de6-a3ae-28b80b60d722-utilities\") pod \"certified-operators-977kk\" (UID: \"92dc9e32-c069-4de6-a3ae-28b80b60d722\") " pod="openshift-marketplace/certified-operators-977kk" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.237932 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjxn6\" (UniqueName: \"kubernetes.io/projected/ee507b1a-9072-40b2-8450-c57209c10a39-kube-api-access-qjxn6\") pod \"community-operators-zx9lt\" (UID: \"ee507b1a-9072-40b2-8450-c57209c10a39\") " pod="openshift-marketplace/community-operators-zx9lt" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.238047 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw586\" (UniqueName: \"kubernetes.io/projected/92dc9e32-c069-4de6-a3ae-28b80b60d722-kube-api-access-vw586\") pod \"certified-operators-977kk\" (UID: \"92dc9e32-c069-4de6-a3ae-28b80b60d722\") " pod="openshift-marketplace/certified-operators-977kk" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.238099 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee507b1a-9072-40b2-8450-c57209c10a39-catalog-content\") pod \"community-operators-zx9lt\" (UID: \"ee507b1a-9072-40b2-8450-c57209c10a39\") " pod="openshift-marketplace/community-operators-zx9lt" Feb 14 11:12:38 crc kubenswrapper[4904]: E0214 11:12:38.238269 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:38.738209355 +0000 UTC m=+149.550974016 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.238873 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92dc9e32-c069-4de6-a3ae-28b80b60d722-catalog-content\") pod \"certified-operators-977kk\" (UID: \"92dc9e32-c069-4de6-a3ae-28b80b60d722\") " pod="openshift-marketplace/certified-operators-977kk" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.239245 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92dc9e32-c069-4de6-a3ae-28b80b60d722-utilities\") pod \"certified-operators-977kk\" (UID: \"92dc9e32-c069-4de6-a3ae-28b80b60d722\") " pod="openshift-marketplace/certified-operators-977kk" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.248133 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hfw59" podStartSLOduration=127.248115197 podStartE2EDuration="2m7.248115197s" podCreationTimestamp="2026-02-14 11:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:38.246238407 +0000 UTC m=+149.059003088" watchObservedRunningTime="2026-02-14 11:12:38.248115197 +0000 UTC m=+149.060879858" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.255127 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.307231 4904 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-bfn42 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.37:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.307289 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" podUID="e831e54b-3826-4415-9b43-ce02e5ed81bf" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.37:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.308235 4904 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-4kpjx container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.308264 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" podUID="345bb34c-a9b2-4e88-94ce-3187b8933057" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.346183 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-fjbmw" podStartSLOduration=11.346169406 podStartE2EDuration="11.346169406s" podCreationTimestamp="2026-02-14 11:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:38.343999459 +0000 UTC m=+149.156764120" watchObservedRunningTime="2026-02-14 11:12:38.346169406 +0000 UTC m=+149.158934067" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.351484 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.351526 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee507b1a-9072-40b2-8450-c57209c10a39-utilities\") pod \"community-operators-zx9lt\" (UID: \"ee507b1a-9072-40b2-8450-c57209c10a39\") " pod="openshift-marketplace/community-operators-zx9lt" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.351584 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjxn6\" (UniqueName: \"kubernetes.io/projected/ee507b1a-9072-40b2-8450-c57209c10a39-kube-api-access-qjxn6\") pod \"community-operators-zx9lt\" (UID: \"ee507b1a-9072-40b2-8450-c57209c10a39\") " pod="openshift-marketplace/community-operators-zx9lt" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.351646 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee507b1a-9072-40b2-8450-c57209c10a39-catalog-content\") pod \"community-operators-zx9lt\" (UID: \"ee507b1a-9072-40b2-8450-c57209c10a39\") " pod="openshift-marketplace/community-operators-zx9lt" Feb 14 11:12:38 crc kubenswrapper[4904]: E0214 11:12:38.351858 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:38.851787345 +0000 UTC m=+149.664552006 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.352189 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee507b1a-9072-40b2-8450-c57209c10a39-catalog-content\") pod \"community-operators-zx9lt\" (UID: \"ee507b1a-9072-40b2-8450-c57209c10a39\") " pod="openshift-marketplace/community-operators-zx9lt" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.352580 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee507b1a-9072-40b2-8450-c57209c10a39-utilities\") pod \"community-operators-zx9lt\" (UID: \"ee507b1a-9072-40b2-8450-c57209c10a39\") " pod="openshift-marketplace/community-operators-zx9lt" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.352600 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw586\" (UniqueName: \"kubernetes.io/projected/92dc9e32-c069-4de6-a3ae-28b80b60d722-kube-api-access-vw586\") pod \"certified-operators-977kk\" (UID: \"92dc9e32-c069-4de6-a3ae-28b80b60d722\") " pod="openshift-marketplace/certified-operators-977kk" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.373116 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qhldd" event={"ID":"966787c0-c9ad-460e-be70-26cb6acbcf0f","Type":"ContainerStarted","Data":"2c01b2378a4d95dca74f267d116e1fbac5ba0b1b5fdff4c641de81755ce931a3"} Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.376336 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wlsqb"] Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.377247 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlsqb" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.397652 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjxn6\" (UniqueName: \"kubernetes.io/projected/ee507b1a-9072-40b2-8450-c57209c10a39-kube-api-access-qjxn6\") pod \"community-operators-zx9lt\" (UID: \"ee507b1a-9072-40b2-8450-c57209c10a39\") " pod="openshift-marketplace/community-operators-zx9lt" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.398900 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6e37b832-8ee6-4eec-b048-3517ae74ed77","Type":"ContainerStarted","Data":"54cbe1ff9cefede84f623239135bb5ca6cfa736c741659c849a646aaa3dfbfd2"} Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.398934 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-fjbmw" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.410814 4904 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9pzkg container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.410877 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" podUID="f130a8a5-24b3-4916-a0c7-b1fa5e55cad4" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.411026 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wlsqb"] Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.415272 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-977kk" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.454159 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.454537 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28f09505-e5c3-4105-88d5-9a6d5ed033e9-utilities\") pod \"certified-operators-wlsqb\" (UID: \"28f09505-e5c3-4105-88d5-9a6d5ed033e9\") " pod="openshift-marketplace/certified-operators-wlsqb" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.454603 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5x2x\" (UniqueName: \"kubernetes.io/projected/28f09505-e5c3-4105-88d5-9a6d5ed033e9-kube-api-access-h5x2x\") pod \"certified-operators-wlsqb\" (UID: \"28f09505-e5c3-4105-88d5-9a6d5ed033e9\") " pod="openshift-marketplace/certified-operators-wlsqb" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.454679 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28f09505-e5c3-4105-88d5-9a6d5ed033e9-catalog-content\") pod \"certified-operators-wlsqb\" (UID: \"28f09505-e5c3-4105-88d5-9a6d5ed033e9\") " pod="openshift-marketplace/certified-operators-wlsqb" Feb 14 11:12:38 crc kubenswrapper[4904]: E0214 11:12:38.454798 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:38.954783865 +0000 UTC m=+149.767548526 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.537760 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx9lt" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.555922 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28f09505-e5c3-4105-88d5-9a6d5ed033e9-utilities\") pod \"certified-operators-wlsqb\" (UID: \"28f09505-e5c3-4105-88d5-9a6d5ed033e9\") " pod="openshift-marketplace/certified-operators-wlsqb" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.555966 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.555991 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5x2x\" (UniqueName: \"kubernetes.io/projected/28f09505-e5c3-4105-88d5-9a6d5ed033e9-kube-api-access-h5x2x\") pod \"certified-operators-wlsqb\" (UID: \"28f09505-e5c3-4105-88d5-9a6d5ed033e9\") " pod="openshift-marketplace/certified-operators-wlsqb" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.556025 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28f09505-e5c3-4105-88d5-9a6d5ed033e9-catalog-content\") pod \"certified-operators-wlsqb\" (UID: \"28f09505-e5c3-4105-88d5-9a6d5ed033e9\") " pod="openshift-marketplace/certified-operators-wlsqb" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.556449 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28f09505-e5c3-4105-88d5-9a6d5ed033e9-utilities\") pod \"certified-operators-wlsqb\" (UID: \"28f09505-e5c3-4105-88d5-9a6d5ed033e9\") " pod="openshift-marketplace/certified-operators-wlsqb" Feb 14 11:12:38 crc kubenswrapper[4904]: E0214 11:12:38.556743 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:39.056730657 +0000 UTC m=+149.869495318 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.560160 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28f09505-e5c3-4105-88d5-9a6d5ed033e9-catalog-content\") pod \"certified-operators-wlsqb\" (UID: \"28f09505-e5c3-4105-88d5-9a6d5ed033e9\") " pod="openshift-marketplace/certified-operators-wlsqb" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.611260 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5x2x\" (UniqueName: \"kubernetes.io/projected/28f09505-e5c3-4105-88d5-9a6d5ed033e9-kube-api-access-h5x2x\") pod \"certified-operators-wlsqb\" (UID: \"28f09505-e5c3-4105-88d5-9a6d5ed033e9\") " pod="openshift-marketplace/certified-operators-wlsqb" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.659517 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:38 crc kubenswrapper[4904]: E0214 11:12:38.659870 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:39.15985369 +0000 UTC m=+149.972618351 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.733147 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.733127406 podStartE2EDuration="3.733127406s" podCreationTimestamp="2026-02-14 11:12:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:38.683220947 +0000 UTC m=+149.495985608" watchObservedRunningTime="2026-02-14 11:12:38.733127406 +0000 UTC m=+149.545892067" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.736885 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlsqb" Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.777623 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:38 crc kubenswrapper[4904]: E0214 11:12:38.778524 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:39.278512644 +0000 UTC m=+150.091277305 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.882993 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:38 crc kubenswrapper[4904]: E0214 11:12:38.883458 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:39.383443845 +0000 UTC m=+150.196208506 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:38 crc kubenswrapper[4904]: I0214 11:12:38.984555 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:38 crc kubenswrapper[4904]: E0214 11:12:38.984957 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:39.484937265 +0000 UTC m=+150.297701926 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.063631 4904 patch_prober.go:28] interesting pod/router-default-5444994796-qh6sr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 14 11:12:39 crc kubenswrapper[4904]: [-]has-synced failed: reason withheld Feb 14 11:12:39 crc kubenswrapper[4904]: [+]process-running ok Feb 14 11:12:39 crc kubenswrapper[4904]: healthz check failed Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.063685 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qh6sr" podUID="91b7cb2e-1781-41da-9c90-13b40be0a2b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.086660 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:39 crc kubenswrapper[4904]: E0214 11:12:39.087408 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:39.587392151 +0000 UTC m=+150.400156812 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.188404 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:39 crc kubenswrapper[4904]: E0214 11:12:39.188707 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:39.688692976 +0000 UTC m=+150.501457637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.289304 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:39 crc kubenswrapper[4904]: E0214 11:12:39.289496 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:39.789468708 +0000 UTC m=+150.602233369 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.289586 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:39 crc kubenswrapper[4904]: E0214 11:12:39.289949 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:39.78994084 +0000 UTC m=+150.602705491 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.390483 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:39 crc kubenswrapper[4904]: E0214 11:12:39.390786 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:39.890765642 +0000 UTC m=+150.703530303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.391045 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:39 crc kubenswrapper[4904]: E0214 11:12:39.391359 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:39.891349499 +0000 UTC m=+150.704114160 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.401513 4904 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-4kpjx container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.401572 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" podUID="345bb34c-a9b2-4e88-94ce-3187b8933057" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.440614 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qhldd" event={"ID":"966787c0-c9ad-460e-be70-26cb6acbcf0f","Type":"ContainerStarted","Data":"506d13675fecdc4372cd860e7625bd039d663181266370ab376644b9475ae76e"} Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.451390 4904 generic.go:334] "Generic (PLEG): container finished" podID="6e37b832-8ee6-4eec-b048-3517ae74ed77" containerID="54cbe1ff9cefede84f623239135bb5ca6cfa736c741659c849a646aaa3dfbfd2" exitCode=0 Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.451451 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6e37b832-8ee6-4eec-b048-3517ae74ed77","Type":"ContainerDied","Data":"54cbe1ff9cefede84f623239135bb5ca6cfa736c741659c849a646aaa3dfbfd2"} Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.459587 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d9116b93131f5c5af9ca4ca71e2c9a6d9dc088a498de02a074ac01873f70e44d"} Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.487953 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.488334 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.493259 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:39 crc kubenswrapper[4904]: E0214 11:12:39.493601 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:39.993586318 +0000 UTC m=+150.806350979 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.545478 4904 patch_prober.go:28] interesting pod/apiserver-76f77b778f-t8ggt container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 14 11:12:39 crc kubenswrapper[4904]: [+]log ok Feb 14 11:12:39 crc kubenswrapper[4904]: [+]etcd ok Feb 14 11:12:39 crc kubenswrapper[4904]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 14 11:12:39 crc kubenswrapper[4904]: [+]poststarthook/generic-apiserver-start-informers ok Feb 14 11:12:39 crc kubenswrapper[4904]: [+]poststarthook/max-in-flight-filter ok Feb 14 11:12:39 crc kubenswrapper[4904]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 14 11:12:39 crc kubenswrapper[4904]: [+]poststarthook/image.openshift.io-apiserver-caches ok Feb 14 11:12:39 crc kubenswrapper[4904]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Feb 14 11:12:39 crc kubenswrapper[4904]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Feb 14 11:12:39 crc kubenswrapper[4904]: [+]poststarthook/project.openshift.io-projectcache ok Feb 14 11:12:39 crc kubenswrapper[4904]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Feb 14 11:12:39 crc kubenswrapper[4904]: [+]poststarthook/openshift.io-startinformers ok Feb 14 11:12:39 crc kubenswrapper[4904]: [+]poststarthook/openshift.io-restmapperupdater ok Feb 14 11:12:39 crc kubenswrapper[4904]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 14 11:12:39 crc kubenswrapper[4904]: livez check failed Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.545543 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" podUID="80b9d666-fa50-4d49-8c3f-efbd9d1247a9" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.595074 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:39 crc kubenswrapper[4904]: E0214 11:12:39.596877 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:40.096866225 +0000 UTC m=+150.909630886 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.700710 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:39 crc kubenswrapper[4904]: E0214 11:12:39.701154 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:40.201138839 +0000 UTC m=+151.013903500 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.763371 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.764070 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.785304 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.802635 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:39 crc kubenswrapper[4904]: E0214 11:12:39.802943 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:40.302932328 +0000 UTC m=+151.115696989 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.848370 4904 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-4kpjx container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.848418 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" podUID="345bb34c-a9b2-4e88-94ce-3187b8933057" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 14 11:12:39 crc kubenswrapper[4904]: W0214 11:12:39.869057 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-d4cbff787151e8d629f1114feec9ab4e624240456a773e6e9d012ffd9db48102 WatchSource:0}: Error finding container d4cbff787151e8d629f1114feec9ab4e624240456a773e6e9d012ffd9db48102: Status 404 returned error can't find the container with id d4cbff787151e8d629f1114feec9ab4e624240456a773e6e9d012ffd9db48102 Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.904072 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:39 crc kubenswrapper[4904]: E0214 11:12:39.905096 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:40.405081676 +0000 UTC m=+151.217846327 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.906168 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4kpjx" Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.928200 4904 patch_prober.go:28] interesting pod/downloads-7954f5f757-bzps5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.928241 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bzps5" podUID="3cb9bb68-546e-4e52-9a15-5aeee129a5b4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.928519 4904 patch_prober.go:28] interesting pod/downloads-7954f5f757-bzps5 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.928534 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-bzps5" podUID="3cb9bb68-546e-4e52-9a15-5aeee129a5b4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.955116 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nvpkp"] Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.956179 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nvpkp" Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.958011 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 14 11:12:39 crc kubenswrapper[4904]: I0214 11:12:39.978254 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nvpkp"] Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.007626 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:40 crc kubenswrapper[4904]: E0214 11:12:40.009274 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:40.509262426 +0000 UTC m=+151.322027087 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.038964 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bvpm5"] Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.040623 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zx9lt"] Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.072630 4904 patch_prober.go:28] interesting pod/router-default-5444994796-qh6sr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 14 11:12:40 crc kubenswrapper[4904]: [-]has-synced failed: reason withheld Feb 14 11:12:40 crc kubenswrapper[4904]: [+]process-running ok Feb 14 11:12:40 crc kubenswrapper[4904]: healthz check failed Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.072676 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qh6sr" podUID="91b7cb2e-1781-41da-9c90-13b40be0a2b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.109568 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.109722 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94b8c7dc-4ce9-4434-a387-e19139fe8a3d-catalog-content\") pod \"redhat-marketplace-nvpkp\" (UID: \"94b8c7dc-4ce9-4434-a387-e19139fe8a3d\") " pod="openshift-marketplace/redhat-marketplace-nvpkp" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.109743 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdhkw\" (UniqueName: \"kubernetes.io/projected/94b8c7dc-4ce9-4434-a387-e19139fe8a3d-kube-api-access-rdhkw\") pod \"redhat-marketplace-nvpkp\" (UID: \"94b8c7dc-4ce9-4434-a387-e19139fe8a3d\") " pod="openshift-marketplace/redhat-marketplace-nvpkp" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.109783 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94b8c7dc-4ce9-4434-a387-e19139fe8a3d-utilities\") pod \"redhat-marketplace-nvpkp\" (UID: \"94b8c7dc-4ce9-4434-a387-e19139fe8a3d\") " pod="openshift-marketplace/redhat-marketplace-nvpkp" Feb 14 11:12:40 crc kubenswrapper[4904]: E0214 11:12:40.110680 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:40.610665084 +0000 UTC m=+151.423429745 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.210913 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94b8c7dc-4ce9-4434-a387-e19139fe8a3d-catalog-content\") pod \"redhat-marketplace-nvpkp\" (UID: \"94b8c7dc-4ce9-4434-a387-e19139fe8a3d\") " pod="openshift-marketplace/redhat-marketplace-nvpkp" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.211205 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdhkw\" (UniqueName: \"kubernetes.io/projected/94b8c7dc-4ce9-4434-a387-e19139fe8a3d-kube-api-access-rdhkw\") pod \"redhat-marketplace-nvpkp\" (UID: \"94b8c7dc-4ce9-4434-a387-e19139fe8a3d\") " pod="openshift-marketplace/redhat-marketplace-nvpkp" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.211233 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.211252 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94b8c7dc-4ce9-4434-a387-e19139fe8a3d-utilities\") pod \"redhat-marketplace-nvpkp\" (UID: \"94b8c7dc-4ce9-4434-a387-e19139fe8a3d\") " pod="openshift-marketplace/redhat-marketplace-nvpkp" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.211611 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94b8c7dc-4ce9-4434-a387-e19139fe8a3d-utilities\") pod \"redhat-marketplace-nvpkp\" (UID: \"94b8c7dc-4ce9-4434-a387-e19139fe8a3d\") " pod="openshift-marketplace/redhat-marketplace-nvpkp" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.211827 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94b8c7dc-4ce9-4434-a387-e19139fe8a3d-catalog-content\") pod \"redhat-marketplace-nvpkp\" (UID: \"94b8c7dc-4ce9-4434-a387-e19139fe8a3d\") " pod="openshift-marketplace/redhat-marketplace-nvpkp" Feb 14 11:12:40 crc kubenswrapper[4904]: E0214 11:12:40.212280 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:40.712268918 +0000 UTC m=+151.525033579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.295921 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdhkw\" (UniqueName: \"kubernetes.io/projected/94b8c7dc-4ce9-4434-a387-e19139fe8a3d-kube-api-access-rdhkw\") pod \"redhat-marketplace-nvpkp\" (UID: \"94b8c7dc-4ce9-4434-a387-e19139fe8a3d\") " pod="openshift-marketplace/redhat-marketplace-nvpkp" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.296626 4904 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.302999 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-977kk"] Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.314384 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:40 crc kubenswrapper[4904]: E0214 11:12:40.314747 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:40.814728303 +0000 UTC m=+151.627492964 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.325879 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nvpkp" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.340067 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-br8st"] Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.341204 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-br8st" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.352376 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wlsqb"] Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.363117 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-br8st"] Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.423540 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5692d85-aa82-486b-889e-a8964e50b538-utilities\") pod \"redhat-marketplace-br8st\" (UID: \"f5692d85-aa82-486b-889e-a8964e50b538\") " pod="openshift-marketplace/redhat-marketplace-br8st" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.423645 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kf99\" (UniqueName: \"kubernetes.io/projected/f5692d85-aa82-486b-889e-a8964e50b538-kube-api-access-2kf99\") pod \"redhat-marketplace-br8st\" (UID: \"f5692d85-aa82-486b-889e-a8964e50b538\") " pod="openshift-marketplace/redhat-marketplace-br8st" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.423684 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.423703 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5692d85-aa82-486b-889e-a8964e50b538-catalog-content\") pod \"redhat-marketplace-br8st\" (UID: \"f5692d85-aa82-486b-889e-a8964e50b538\") " pod="openshift-marketplace/redhat-marketplace-br8st" Feb 14 11:12:40 crc kubenswrapper[4904]: E0214 11:12:40.424014 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:40.924000779 +0000 UTC m=+151.736765440 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.463313 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-977kk" event={"ID":"92dc9e32-c069-4de6-a3ae-28b80b60d722","Type":"ContainerStarted","Data":"fecbde0a237da81f7630e72ce2d0a2dc763ad16d5ac2f9bbe555882218609cfe"} Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.464483 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7ace167cda8f6ad07cfb838348af6f98d6d5480afd70616e824a22b02ec5bdab"} Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.467126 4904 generic.go:334] "Generic (PLEG): container finished" podID="ee507b1a-9072-40b2-8450-c57209c10a39" containerID="2273b2c80df3f5550f5b4463f68201663f6da1d8dff7a85cd8258b45fb817db7" exitCode=0 Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.467182 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx9lt" event={"ID":"ee507b1a-9072-40b2-8450-c57209c10a39","Type":"ContainerDied","Data":"2273b2c80df3f5550f5b4463f68201663f6da1d8dff7a85cd8258b45fb817db7"} Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.467199 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx9lt" event={"ID":"ee507b1a-9072-40b2-8450-c57209c10a39","Type":"ContainerStarted","Data":"e5a24a82a2b61dd4adef5d9a31562decd67cbd2548dd8575d54907cddd642297"} Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.468810 4904 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.469665 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"b2d6242ce6ec0597f930a7f17edab638c9e5c709dcece947d8b548ec4d391bff"} Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.469689 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"d4cbff787151e8d629f1114feec9ab4e624240456a773e6e9d012ffd9db48102"} Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.470228 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.474015 4904 generic.go:334] "Generic (PLEG): container finished" podID="63c9ce1a-01e3-42f3-8027-bfda9be0b3ff" containerID="3b8956f5878119fde34b8d263a61fdaa2af4aa0cefe3aabd7f94eea092a181f3" exitCode=0 Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.474290 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bvpm5" event={"ID":"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff","Type":"ContainerDied","Data":"3b8956f5878119fde34b8d263a61fdaa2af4aa0cefe3aabd7f94eea092a181f3"} Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.474321 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bvpm5" event={"ID":"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff","Type":"ContainerStarted","Data":"a9cfcde1152608ae1472a0761cda82a2bf852954afbcd565db74b202ecf6632b"} Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.491451 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qhldd" event={"ID":"966787c0-c9ad-460e-be70-26cb6acbcf0f","Type":"ContainerStarted","Data":"cd1118a84681b501a5504dbee55782f3cfb1ecb1f3ec279982a19af9cb92c98e"} Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.507640 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlsqb" event={"ID":"28f09505-e5c3-4105-88d5-9a6d5ed033e9","Type":"ContainerStarted","Data":"c68ee9c351a9fa1a2d72ec1053e093506230ba6c2190e3e478573e851c2ae420"} Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.509251 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"da235e5bde211768d6ff708438a8fa5570d2a9f4d4bdc12e8e38c54edb29826e"} Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.514716 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.524502 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.524708 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kf99\" (UniqueName: \"kubernetes.io/projected/f5692d85-aa82-486b-889e-a8964e50b538-kube-api-access-2kf99\") pod \"redhat-marketplace-br8st\" (UID: \"f5692d85-aa82-486b-889e-a8964e50b538\") " pod="openshift-marketplace/redhat-marketplace-br8st" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.524746 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5692d85-aa82-486b-889e-a8964e50b538-catalog-content\") pod \"redhat-marketplace-br8st\" (UID: \"f5692d85-aa82-486b-889e-a8964e50b538\") " pod="openshift-marketplace/redhat-marketplace-br8st" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.524783 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5692d85-aa82-486b-889e-a8964e50b538-utilities\") pod \"redhat-marketplace-br8st\" (UID: \"f5692d85-aa82-486b-889e-a8964e50b538\") " pod="openshift-marketplace/redhat-marketplace-br8st" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.526137 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5692d85-aa82-486b-889e-a8964e50b538-utilities\") pod \"redhat-marketplace-br8st\" (UID: \"f5692d85-aa82-486b-889e-a8964e50b538\") " pod="openshift-marketplace/redhat-marketplace-br8st" Feb 14 11:12:40 crc kubenswrapper[4904]: E0214 11:12:40.526302 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:41.02628074 +0000 UTC m=+151.839045401 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.526340 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5692d85-aa82-486b-889e-a8964e50b538-catalog-content\") pod \"redhat-marketplace-br8st\" (UID: \"f5692d85-aa82-486b-889e-a8964e50b538\") " pod="openshift-marketplace/redhat-marketplace-br8st" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.526419 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-jrpqw" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.544497 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-qhldd" podStartSLOduration=13.544481771 podStartE2EDuration="13.544481771s" podCreationTimestamp="2026-02-14 11:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:40.5153021 +0000 UTC m=+151.328066761" watchObservedRunningTime="2026-02-14 11:12:40.544481771 +0000 UTC m=+151.357246432" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.561695 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kf99\" (UniqueName: \"kubernetes.io/projected/f5692d85-aa82-486b-889e-a8964e50b538-kube-api-access-2kf99\") pod \"redhat-marketplace-br8st\" (UID: \"f5692d85-aa82-486b-889e-a8964e50b538\") " pod="openshift-marketplace/redhat-marketplace-br8st" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.629566 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:40 crc kubenswrapper[4904]: E0214 11:12:40.636706 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:41.136692046 +0000 UTC m=+151.949456707 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.672229 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-br8st" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.735774 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:40 crc kubenswrapper[4904]: E0214 11:12:40.736627 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:41.236612175 +0000 UTC m=+152.049376826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.818645 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.818705 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.838270 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:40 crc kubenswrapper[4904]: E0214 11:12:40.838547 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-14 11:12:41.338535106 +0000 UTC m=+152.151299767 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbh4h" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.843221 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.845523 4904 patch_prober.go:28] interesting pod/console-f9d7485db-7c6cp container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.36:8443/health\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.845579 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-7c6cp" podUID="261ad318-57b3-48ca-8dac-e8bd53528694" containerName="console" probeResult="failure" output="Get \"https://10.217.0.36:8443/health\": dial tcp 10.217.0.36:8443: connect: connection refused" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.876039 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nvpkp"] Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.938868 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:40 crc kubenswrapper[4904]: E0214 11:12:40.940036 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-14 11:12:41.440015867 +0000 UTC m=+152.252780528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.951114 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-krh2c"] Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.952579 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krh2c" Feb 14 11:12:40 crc kubenswrapper[4904]: I0214 11:12:40.956675 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.006706 4904 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-14T11:12:40.296642825Z","Handler":null,"Name":""} Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.021916 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-krh2c"] Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.051087 4904 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.051138 4904 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.051613 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49e91378-b5ff-4480-abbf-765ccc061978-utilities\") pod \"redhat-operators-krh2c\" (UID: \"49e91378-b5ff-4480-abbf-765ccc061978\") " pod="openshift-marketplace/redhat-operators-krh2c" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.051643 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv727\" (UniqueName: \"kubernetes.io/projected/49e91378-b5ff-4480-abbf-765ccc061978-kube-api-access-pv727\") pod \"redhat-operators-krh2c\" (UID: \"49e91378-b5ff-4480-abbf-765ccc061978\") " pod="openshift-marketplace/redhat-operators-krh2c" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.051674 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49e91378-b5ff-4480-abbf-765ccc061978-catalog-content\") pod \"redhat-operators-krh2c\" (UID: \"49e91378-b5ff-4480-abbf-765ccc061978\") " pod="openshift-marketplace/redhat-operators-krh2c" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.051712 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.056942 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:41 crc kubenswrapper[4904]: E0214 11:12:41.077041 4904 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28f09505_e5c3_4105_88d5_9a6d5ed033e9.slice/crio-f9c8683559f28895e271f2eefc21b2d9b9350ccea0cb3b9f4b127458bf1cae33.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92dc9e32_c069_4de6_a3ae_28b80b60d722.slice/crio-22e81e28a495ab6f07d67874f918993f4cc3d8586268e06e326f31c954e5e63b.scope\": RecentStats: unable to find data in memory cache]" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.083606 4904 patch_prober.go:28] interesting pod/router-default-5444994796-qh6sr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 14 11:12:41 crc kubenswrapper[4904]: [-]has-synced failed: reason withheld Feb 14 11:12:41 crc kubenswrapper[4904]: [+]process-running ok Feb 14 11:12:41 crc kubenswrapper[4904]: healthz check failed Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.083653 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qh6sr" podUID="91b7cb2e-1781-41da-9c90-13b40be0a2b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.085818 4904 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.085878 4904 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.161738 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49e91378-b5ff-4480-abbf-765ccc061978-utilities\") pod \"redhat-operators-krh2c\" (UID: \"49e91378-b5ff-4480-abbf-765ccc061978\") " pod="openshift-marketplace/redhat-operators-krh2c" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.161778 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pv727\" (UniqueName: \"kubernetes.io/projected/49e91378-b5ff-4480-abbf-765ccc061978-kube-api-access-pv727\") pod \"redhat-operators-krh2c\" (UID: \"49e91378-b5ff-4480-abbf-765ccc061978\") " pod="openshift-marketplace/redhat-operators-krh2c" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.161850 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49e91378-b5ff-4480-abbf-765ccc061978-catalog-content\") pod \"redhat-operators-krh2c\" (UID: \"49e91378-b5ff-4480-abbf-765ccc061978\") " pod="openshift-marketplace/redhat-operators-krh2c" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.163151 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49e91378-b5ff-4480-abbf-765ccc061978-catalog-content\") pod \"redhat-operators-krh2c\" (UID: \"49e91378-b5ff-4480-abbf-765ccc061978\") " pod="openshift-marketplace/redhat-operators-krh2c" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.163254 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49e91378-b5ff-4480-abbf-765ccc061978-utilities\") pod \"redhat-operators-krh2c\" (UID: \"49e91378-b5ff-4480-abbf-765ccc061978\") " pod="openshift-marketplace/redhat-operators-krh2c" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.255351 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pv727\" (UniqueName: \"kubernetes.io/projected/49e91378-b5ff-4480-abbf-765ccc061978-kube-api-access-pv727\") pod \"redhat-operators-krh2c\" (UID: \"49e91378-b5ff-4480-abbf-765ccc061978\") " pod="openshift-marketplace/redhat-operators-krh2c" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.360421 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.364041 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xphhn"] Feb 14 11:12:41 crc kubenswrapper[4904]: E0214 11:12:41.364281 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e37b832-8ee6-4eec-b048-3517ae74ed77" containerName="pruner" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.364336 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e37b832-8ee6-4eec-b048-3517ae74ed77" containerName="pruner" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.364600 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e37b832-8ee6-4eec-b048-3517ae74ed77" containerName="pruner" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.365788 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xphhn" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.394763 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krh2c" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.404074 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbh4h\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.406891 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xphhn"] Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.481473 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e37b832-8ee6-4eec-b048-3517ae74ed77-kube-api-access\") pod \"6e37b832-8ee6-4eec-b048-3517ae74ed77\" (UID: \"6e37b832-8ee6-4eec-b048-3517ae74ed77\") " Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.481634 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.481676 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6e37b832-8ee6-4eec-b048-3517ae74ed77-kubelet-dir\") pod \"6e37b832-8ee6-4eec-b048-3517ae74ed77\" (UID: \"6e37b832-8ee6-4eec-b048-3517ae74ed77\") " Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.482077 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3c0e6e8-ac12-4225-b17f-095591f8d2a7-catalog-content\") pod \"redhat-operators-xphhn\" (UID: \"e3c0e6e8-ac12-4225-b17f-095591f8d2a7\") " pod="openshift-marketplace/redhat-operators-xphhn" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.482122 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jcv6\" (UniqueName: \"kubernetes.io/projected/e3c0e6e8-ac12-4225-b17f-095591f8d2a7-kube-api-access-8jcv6\") pod \"redhat-operators-xphhn\" (UID: \"e3c0e6e8-ac12-4225-b17f-095591f8d2a7\") " pod="openshift-marketplace/redhat-operators-xphhn" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.482176 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3c0e6e8-ac12-4225-b17f-095591f8d2a7-utilities\") pod \"redhat-operators-xphhn\" (UID: \"e3c0e6e8-ac12-4225-b17f-095591f8d2a7\") " pod="openshift-marketplace/redhat-operators-xphhn" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.482332 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6e37b832-8ee6-4eec-b048-3517ae74ed77-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6e37b832-8ee6-4eec-b048-3517ae74ed77" (UID: "6e37b832-8ee6-4eec-b048-3517ae74ed77"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.498004 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e37b832-8ee6-4eec-b048-3517ae74ed77-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6e37b832-8ee6-4eec-b048-3517ae74ed77" (UID: "6e37b832-8ee6-4eec-b048-3517ae74ed77"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.532461 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6e37b832-8ee6-4eec-b048-3517ae74ed77","Type":"ContainerDied","Data":"ab71ba87990201eb4e7cd757ed766aa61a3b86945930103e35929b849e440296"} Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.532507 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab71ba87990201eb4e7cd757ed766aa61a3b86945930103e35929b849e440296" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.532567 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.557753 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-br8st"] Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.563696 4904 generic.go:334] "Generic (PLEG): container finished" podID="28f09505-e5c3-4105-88d5-9a6d5ed033e9" containerID="f9c8683559f28895e271f2eefc21b2d9b9350ccea0cb3b9f4b127458bf1cae33" exitCode=0 Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.563908 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlsqb" event={"ID":"28f09505-e5c3-4105-88d5-9a6d5ed033e9","Type":"ContainerDied","Data":"f9c8683559f28895e271f2eefc21b2d9b9350ccea0cb3b9f4b127458bf1cae33"} Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.565715 4904 generic.go:334] "Generic (PLEG): container finished" podID="94b8c7dc-4ce9-4434-a387-e19139fe8a3d" containerID="20f6bcba8893436605e59344de145196a5f5ed8e267d59dff7f682eaf65f10ec" exitCode=0 Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.565757 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nvpkp" event={"ID":"94b8c7dc-4ce9-4434-a387-e19139fe8a3d","Type":"ContainerDied","Data":"20f6bcba8893436605e59344de145196a5f5ed8e267d59dff7f682eaf65f10ec"} Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.565778 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nvpkp" event={"ID":"94b8c7dc-4ce9-4434-a387-e19139fe8a3d","Type":"ContainerStarted","Data":"782d6d853e618365d7d2209b9b4d5d1d260fc2606a84afc39e8d55e1473acaa4"} Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.578030 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.583882 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3c0e6e8-ac12-4225-b17f-095591f8d2a7-catalog-content\") pod \"redhat-operators-xphhn\" (UID: \"e3c0e6e8-ac12-4225-b17f-095591f8d2a7\") " pod="openshift-marketplace/redhat-operators-xphhn" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.583916 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jcv6\" (UniqueName: \"kubernetes.io/projected/e3c0e6e8-ac12-4225-b17f-095591f8d2a7-kube-api-access-8jcv6\") pod \"redhat-operators-xphhn\" (UID: \"e3c0e6e8-ac12-4225-b17f-095591f8d2a7\") " pod="openshift-marketplace/redhat-operators-xphhn" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.583957 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3c0e6e8-ac12-4225-b17f-095591f8d2a7-utilities\") pod \"redhat-operators-xphhn\" (UID: \"e3c0e6e8-ac12-4225-b17f-095591f8d2a7\") " pod="openshift-marketplace/redhat-operators-xphhn" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.584005 4904 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6e37b832-8ee6-4eec-b048-3517ae74ed77-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.584016 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e37b832-8ee6-4eec-b048-3517ae74ed77-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.584379 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3c0e6e8-ac12-4225-b17f-095591f8d2a7-utilities\") pod \"redhat-operators-xphhn\" (UID: \"e3c0e6e8-ac12-4225-b17f-095591f8d2a7\") " pod="openshift-marketplace/redhat-operators-xphhn" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.584585 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3c0e6e8-ac12-4225-b17f-095591f8d2a7-catalog-content\") pod \"redhat-operators-xphhn\" (UID: \"e3c0e6e8-ac12-4225-b17f-095591f8d2a7\") " pod="openshift-marketplace/redhat-operators-xphhn" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.585737 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"9ca9654034541e76a22d728f11b07dadd7abe5cdc9e1974845436a436c423ddf"} Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.586462 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.588173 4904 generic.go:334] "Generic (PLEG): container finished" podID="92dc9e32-c069-4de6-a3ae-28b80b60d722" containerID="22e81e28a495ab6f07d67874f918993f4cc3d8586268e06e326f31c954e5e63b" exitCode=0 Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.588216 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-977kk" event={"ID":"92dc9e32-c069-4de6-a3ae-28b80b60d722","Type":"ContainerDied","Data":"22e81e28a495ab6f07d67874f918993f4cc3d8586268e06e326f31c954e5e63b"} Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.603873 4904 generic.go:334] "Generic (PLEG): container finished" podID="6d9f5d3d-fb18-4302-8b6b-62d4b8999b65" containerID="6a6ed2121e9a25309d4d0411dd1abe47ea2f42a715eae146c16d359ff323786c" exitCode=0 Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.604422 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx" event={"ID":"6d9f5d3d-fb18-4302-8b6b-62d4b8999b65","Type":"ContainerDied","Data":"6a6ed2121e9a25309d4d0411dd1abe47ea2f42a715eae146c16d359ff323786c"} Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.619246 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jcv6\" (UniqueName: \"kubernetes.io/projected/e3c0e6e8-ac12-4225-b17f-095591f8d2a7-kube-api-access-8jcv6\") pod \"redhat-operators-xphhn\" (UID: \"e3c0e6e8-ac12-4225-b17f-095591f8d2a7\") " pod="openshift-marketplace/redhat-operators-xphhn" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.722822 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xphhn" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.856998 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 14 11:12:41 crc kubenswrapper[4904]: I0214 11:12:41.942023 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-krh2c"] Feb 14 11:12:42 crc kubenswrapper[4904]: I0214 11:12:42.064324 4904 patch_prober.go:28] interesting pod/router-default-5444994796-qh6sr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 14 11:12:42 crc kubenswrapper[4904]: [-]has-synced failed: reason withheld Feb 14 11:12:42 crc kubenswrapper[4904]: [+]process-running ok Feb 14 11:12:42 crc kubenswrapper[4904]: healthz check failed Feb 14 11:12:42 crc kubenswrapper[4904]: I0214 11:12:42.064585 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qh6sr" podUID="91b7cb2e-1781-41da-9c90-13b40be0a2b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 14 11:12:42 crc kubenswrapper[4904]: I0214 11:12:42.266419 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xphhn"] Feb 14 11:12:42 crc kubenswrapper[4904]: I0214 11:12:42.349244 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zbh4h"] Feb 14 11:12:42 crc kubenswrapper[4904]: I0214 11:12:42.628218 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xphhn" event={"ID":"e3c0e6e8-ac12-4225-b17f-095591f8d2a7","Type":"ContainerStarted","Data":"00750ee463fe5d80a8b5aaf2129de84ec9145448f7c7200c7ddf9797f8b8f6dc"} Feb 14 11:12:42 crc kubenswrapper[4904]: I0214 11:12:42.632458 4904 generic.go:334] "Generic (PLEG): container finished" podID="49e91378-b5ff-4480-abbf-765ccc061978" containerID="33ff4599352ab484361317e1141542245cd3827b024c49d1b281f4433e833f9c" exitCode=0 Feb 14 11:12:42 crc kubenswrapper[4904]: I0214 11:12:42.632502 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krh2c" event={"ID":"49e91378-b5ff-4480-abbf-765ccc061978","Type":"ContainerDied","Data":"33ff4599352ab484361317e1141542245cd3827b024c49d1b281f4433e833f9c"} Feb 14 11:12:42 crc kubenswrapper[4904]: I0214 11:12:42.632521 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krh2c" event={"ID":"49e91378-b5ff-4480-abbf-765ccc061978","Type":"ContainerStarted","Data":"01a2e06d56c934699ea02506921cc080998cfdf2a52e96dd13e3d62107da12e8"} Feb 14 11:12:42 crc kubenswrapper[4904]: I0214 11:12:42.658072 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" event={"ID":"d97a9335-cf15-4b30-ba2e-846e0daed72d","Type":"ContainerStarted","Data":"8d9cdb862c4416609468bf22f5a8fd04c4e1d268f029a75897a9b5cff99c922a"} Feb 14 11:12:42 crc kubenswrapper[4904]: I0214 11:12:42.686154 4904 generic.go:334] "Generic (PLEG): container finished" podID="f5692d85-aa82-486b-889e-a8964e50b538" containerID="9991f3746f32eb38dfbe0661228622068ded07209f3e12a996c2451cf248147c" exitCode=0 Feb 14 11:12:42 crc kubenswrapper[4904]: I0214 11:12:42.687007 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-br8st" event={"ID":"f5692d85-aa82-486b-889e-a8964e50b538","Type":"ContainerDied","Data":"9991f3746f32eb38dfbe0661228622068ded07209f3e12a996c2451cf248147c"} Feb 14 11:12:42 crc kubenswrapper[4904]: I0214 11:12:42.687029 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-br8st" event={"ID":"f5692d85-aa82-486b-889e-a8964e50b538","Type":"ContainerStarted","Data":"6bb6191767325a72c03b131b48bdf7e32e898e35e52da6e26bc10e3abdf80da0"} Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.031959 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx" Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.063188 4904 patch_prober.go:28] interesting pod/router-default-5444994796-qh6sr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 14 11:12:43 crc kubenswrapper[4904]: [-]has-synced failed: reason withheld Feb 14 11:12:43 crc kubenswrapper[4904]: [+]process-running ok Feb 14 11:12:43 crc kubenswrapper[4904]: healthz check failed Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.063256 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qh6sr" podUID="91b7cb2e-1781-41da-9c90-13b40be0a2b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.122888 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwplk\" (UniqueName: \"kubernetes.io/projected/6d9f5d3d-fb18-4302-8b6b-62d4b8999b65-kube-api-access-rwplk\") pod \"6d9f5d3d-fb18-4302-8b6b-62d4b8999b65\" (UID: \"6d9f5d3d-fb18-4302-8b6b-62d4b8999b65\") " Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.123316 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d9f5d3d-fb18-4302-8b6b-62d4b8999b65-secret-volume\") pod \"6d9f5d3d-fb18-4302-8b6b-62d4b8999b65\" (UID: \"6d9f5d3d-fb18-4302-8b6b-62d4b8999b65\") " Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.123341 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d9f5d3d-fb18-4302-8b6b-62d4b8999b65-config-volume\") pod \"6d9f5d3d-fb18-4302-8b6b-62d4b8999b65\" (UID: \"6d9f5d3d-fb18-4302-8b6b-62d4b8999b65\") " Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.124154 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d9f5d3d-fb18-4302-8b6b-62d4b8999b65-config-volume" (OuterVolumeSpecName: "config-volume") pod "6d9f5d3d-fb18-4302-8b6b-62d4b8999b65" (UID: "6d9f5d3d-fb18-4302-8b6b-62d4b8999b65"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.132411 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d9f5d3d-fb18-4302-8b6b-62d4b8999b65-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6d9f5d3d-fb18-4302-8b6b-62d4b8999b65" (UID: "6d9f5d3d-fb18-4302-8b6b-62d4b8999b65"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.135297 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d9f5d3d-fb18-4302-8b6b-62d4b8999b65-kube-api-access-rwplk" (OuterVolumeSpecName: "kube-api-access-rwplk") pod "6d9f5d3d-fb18-4302-8b6b-62d4b8999b65" (UID: "6d9f5d3d-fb18-4302-8b6b-62d4b8999b65"). InnerVolumeSpecName "kube-api-access-rwplk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.224907 4904 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d9f5d3d-fb18-4302-8b6b-62d4b8999b65-config-volume\") on node \"crc\" DevicePath \"\"" Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.225006 4904 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d9f5d3d-fb18-4302-8b6b-62d4b8999b65-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.225018 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwplk\" (UniqueName: \"kubernetes.io/projected/6d9f5d3d-fb18-4302-8b6b-62d4b8999b65-kube-api-access-rwplk\") on node \"crc\" DevicePath \"\"" Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.796557 4904 generic.go:334] "Generic (PLEG): container finished" podID="e3c0e6e8-ac12-4225-b17f-095591f8d2a7" containerID="c9b142bf1889a0c950f580730dbf830044d28ca073d00afc7f104d0ec4342fc0" exitCode=0 Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.796701 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xphhn" event={"ID":"e3c0e6e8-ac12-4225-b17f-095591f8d2a7","Type":"ContainerDied","Data":"c9b142bf1889a0c950f580730dbf830044d28ca073d00afc7f104d0ec4342fc0"} Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.803234 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" event={"ID":"d97a9335-cf15-4b30-ba2e-846e0daed72d","Type":"ContainerStarted","Data":"a50f842bce824e3f5855d72f7e9101d5d8ecedc24956419d4ddc8f2ee345d29b"} Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.803645 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.809395 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx" Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.814478 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx" event={"ID":"6d9f5d3d-fb18-4302-8b6b-62d4b8999b65","Type":"ContainerDied","Data":"cb42d973f41b3aa1e0084c56c98b85cfe2206bc43df0751a876efe821a42fbc2"} Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.814517 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb42d973f41b3aa1e0084c56c98b85cfe2206bc43df0751a876efe821a42fbc2" Feb 14 11:12:43 crc kubenswrapper[4904]: I0214 11:12:43.860500 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" podStartSLOduration=133.860466102 podStartE2EDuration="2m13.860466102s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:43.855663134 +0000 UTC m=+154.668427795" watchObservedRunningTime="2026-02-14 11:12:43.860466102 +0000 UTC m=+154.673230763" Feb 14 11:12:44 crc kubenswrapper[4904]: I0214 11:12:44.060488 4904 patch_prober.go:28] interesting pod/router-default-5444994796-qh6sr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 14 11:12:44 crc kubenswrapper[4904]: [-]has-synced failed: reason withheld Feb 14 11:12:44 crc kubenswrapper[4904]: [+]process-running ok Feb 14 11:12:44 crc kubenswrapper[4904]: healthz check failed Feb 14 11:12:44 crc kubenswrapper[4904]: I0214 11:12:44.060573 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qh6sr" podUID="91b7cb2e-1781-41da-9c90-13b40be0a2b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 14 11:12:44 crc kubenswrapper[4904]: I0214 11:12:44.495590 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:44 crc kubenswrapper[4904]: I0214 11:12:44.499905 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-t8ggt" Feb 14 11:12:45 crc kubenswrapper[4904]: I0214 11:12:45.058702 4904 patch_prober.go:28] interesting pod/router-default-5444994796-qh6sr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 14 11:12:45 crc kubenswrapper[4904]: [-]has-synced failed: reason withheld Feb 14 11:12:45 crc kubenswrapper[4904]: [+]process-running ok Feb 14 11:12:45 crc kubenswrapper[4904]: healthz check failed Feb 14 11:12:45 crc kubenswrapper[4904]: I0214 11:12:45.058979 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qh6sr" podUID="91b7cb2e-1781-41da-9c90-13b40be0a2b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 14 11:12:45 crc kubenswrapper[4904]: I0214 11:12:45.882041 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-fjbmw" Feb 14 11:12:46 crc kubenswrapper[4904]: I0214 11:12:46.069055 4904 patch_prober.go:28] interesting pod/router-default-5444994796-qh6sr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 14 11:12:46 crc kubenswrapper[4904]: [-]has-synced failed: reason withheld Feb 14 11:12:46 crc kubenswrapper[4904]: [+]process-running ok Feb 14 11:12:46 crc kubenswrapper[4904]: healthz check failed Feb 14 11:12:46 crc kubenswrapper[4904]: I0214 11:12:46.069111 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qh6sr" podUID="91b7cb2e-1781-41da-9c90-13b40be0a2b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 14 11:12:46 crc kubenswrapper[4904]: I0214 11:12:46.076433 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 14 11:12:46 crc kubenswrapper[4904]: E0214 11:12:46.076742 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d9f5d3d-fb18-4302-8b6b-62d4b8999b65" containerName="collect-profiles" Feb 14 11:12:46 crc kubenswrapper[4904]: I0214 11:12:46.076761 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d9f5d3d-fb18-4302-8b6b-62d4b8999b65" containerName="collect-profiles" Feb 14 11:12:46 crc kubenswrapper[4904]: I0214 11:12:46.076921 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d9f5d3d-fb18-4302-8b6b-62d4b8999b65" containerName="collect-profiles" Feb 14 11:12:46 crc kubenswrapper[4904]: I0214 11:12:46.077330 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 14 11:12:46 crc kubenswrapper[4904]: I0214 11:12:46.080942 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 14 11:12:46 crc kubenswrapper[4904]: I0214 11:12:46.095872 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 14 11:12:46 crc kubenswrapper[4904]: I0214 11:12:46.096270 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 14 11:12:46 crc kubenswrapper[4904]: I0214 11:12:46.202467 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/617704a5-df59-410d-b6a8-b462453ff137-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"617704a5-df59-410d-b6a8-b462453ff137\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 14 11:12:46 crc kubenswrapper[4904]: I0214 11:12:46.202529 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/617704a5-df59-410d-b6a8-b462453ff137-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"617704a5-df59-410d-b6a8-b462453ff137\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 14 11:12:46 crc kubenswrapper[4904]: I0214 11:12:46.303588 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/617704a5-df59-410d-b6a8-b462453ff137-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"617704a5-df59-410d-b6a8-b462453ff137\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 14 11:12:46 crc kubenswrapper[4904]: I0214 11:12:46.303645 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/617704a5-df59-410d-b6a8-b462453ff137-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"617704a5-df59-410d-b6a8-b462453ff137\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 14 11:12:46 crc kubenswrapper[4904]: I0214 11:12:46.304075 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/617704a5-df59-410d-b6a8-b462453ff137-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"617704a5-df59-410d-b6a8-b462453ff137\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 14 11:12:46 crc kubenswrapper[4904]: I0214 11:12:46.336731 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/617704a5-df59-410d-b6a8-b462453ff137-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"617704a5-df59-410d-b6a8-b462453ff137\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 14 11:12:46 crc kubenswrapper[4904]: I0214 11:12:46.383563 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:12:46 crc kubenswrapper[4904]: I0214 11:12:46.383624 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:12:46 crc kubenswrapper[4904]: I0214 11:12:46.408578 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 14 11:12:46 crc kubenswrapper[4904]: I0214 11:12:46.993093 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 14 11:12:47 crc kubenswrapper[4904]: I0214 11:12:47.065690 4904 patch_prober.go:28] interesting pod/router-default-5444994796-qh6sr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 14 11:12:47 crc kubenswrapper[4904]: [-]has-synced failed: reason withheld Feb 14 11:12:47 crc kubenswrapper[4904]: [+]process-running ok Feb 14 11:12:47 crc kubenswrapper[4904]: healthz check failed Feb 14 11:12:47 crc kubenswrapper[4904]: I0214 11:12:47.065778 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qh6sr" podUID="91b7cb2e-1781-41da-9c90-13b40be0a2b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 14 11:12:47 crc kubenswrapper[4904]: I0214 11:12:47.924299 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"617704a5-df59-410d-b6a8-b462453ff137","Type":"ContainerStarted","Data":"7bea0c851dfc564ddf532712a8fe82264af4c70c9a3740b8174c5f28437427c5"} Feb 14 11:12:48 crc kubenswrapper[4904]: I0214 11:12:48.062329 4904 patch_prober.go:28] interesting pod/router-default-5444994796-qh6sr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 14 11:12:48 crc kubenswrapper[4904]: [-]has-synced failed: reason withheld Feb 14 11:12:48 crc kubenswrapper[4904]: [+]process-running ok Feb 14 11:12:48 crc kubenswrapper[4904]: healthz check failed Feb 14 11:12:48 crc kubenswrapper[4904]: I0214 11:12:48.062391 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qh6sr" podUID="91b7cb2e-1781-41da-9c90-13b40be0a2b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 14 11:12:48 crc kubenswrapper[4904]: I0214 11:12:48.904770 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"617704a5-df59-410d-b6a8-b462453ff137","Type":"ContainerStarted","Data":"53239610bf56d99a57eb6450dfa8094aafffb09e37e40492766d8e40f2461f9e"} Feb 14 11:12:48 crc kubenswrapper[4904]: I0214 11:12:48.927728 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.9277072 podStartE2EDuration="2.9277072s" podCreationTimestamp="2026-02-14 11:12:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:12:48.919406281 +0000 UTC m=+159.732170942" watchObservedRunningTime="2026-02-14 11:12:48.9277072 +0000 UTC m=+159.740471861" Feb 14 11:12:49 crc kubenswrapper[4904]: I0214 11:12:49.059435 4904 patch_prober.go:28] interesting pod/router-default-5444994796-qh6sr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 14 11:12:49 crc kubenswrapper[4904]: [-]has-synced failed: reason withheld Feb 14 11:12:49 crc kubenswrapper[4904]: [+]process-running ok Feb 14 11:12:49 crc kubenswrapper[4904]: healthz check failed Feb 14 11:12:49 crc kubenswrapper[4904]: I0214 11:12:49.059533 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qh6sr" podUID="91b7cb2e-1781-41da-9c90-13b40be0a2b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 14 11:12:49 crc kubenswrapper[4904]: I0214 11:12:49.927622 4904 patch_prober.go:28] interesting pod/downloads-7954f5f757-bzps5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Feb 14 11:12:49 crc kubenswrapper[4904]: I0214 11:12:49.927716 4904 patch_prober.go:28] interesting pod/downloads-7954f5f757-bzps5 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Feb 14 11:12:49 crc kubenswrapper[4904]: I0214 11:12:49.930221 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-bzps5" podUID="3cb9bb68-546e-4e52-9a15-5aeee129a5b4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Feb 14 11:12:49 crc kubenswrapper[4904]: I0214 11:12:49.927930 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bzps5" podUID="3cb9bb68-546e-4e52-9a15-5aeee129a5b4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Feb 14 11:12:49 crc kubenswrapper[4904]: I0214 11:12:49.953514 4904 generic.go:334] "Generic (PLEG): container finished" podID="617704a5-df59-410d-b6a8-b462453ff137" containerID="53239610bf56d99a57eb6450dfa8094aafffb09e37e40492766d8e40f2461f9e" exitCode=0 Feb 14 11:12:49 crc kubenswrapper[4904]: I0214 11:12:49.953557 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"617704a5-df59-410d-b6a8-b462453ff137","Type":"ContainerDied","Data":"53239610bf56d99a57eb6450dfa8094aafffb09e37e40492766d8e40f2461f9e"} Feb 14 11:12:50 crc kubenswrapper[4904]: I0214 11:12:50.061690 4904 patch_prober.go:28] interesting pod/router-default-5444994796-qh6sr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 14 11:12:50 crc kubenswrapper[4904]: [-]has-synced failed: reason withheld Feb 14 11:12:50 crc kubenswrapper[4904]: [+]process-running ok Feb 14 11:12:50 crc kubenswrapper[4904]: healthz check failed Feb 14 11:12:50 crc kubenswrapper[4904]: I0214 11:12:50.061740 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qh6sr" podUID="91b7cb2e-1781-41da-9c90-13b40be0a2b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 14 11:12:50 crc kubenswrapper[4904]: I0214 11:12:50.818448 4904 patch_prober.go:28] interesting pod/console-f9d7485db-7c6cp container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.36:8443/health\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Feb 14 11:12:50 crc kubenswrapper[4904]: I0214 11:12:50.818530 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-7c6cp" podUID="261ad318-57b3-48ca-8dac-e8bd53528694" containerName="console" probeResult="failure" output="Get \"https://10.217.0.36:8443/health\": dial tcp 10.217.0.36:8443: connect: connection refused" Feb 14 11:12:51 crc kubenswrapper[4904]: I0214 11:12:51.062198 4904 patch_prober.go:28] interesting pod/router-default-5444994796-qh6sr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 14 11:12:51 crc kubenswrapper[4904]: [-]has-synced failed: reason withheld Feb 14 11:12:51 crc kubenswrapper[4904]: [+]process-running ok Feb 14 11:12:51 crc kubenswrapper[4904]: healthz check failed Feb 14 11:12:51 crc kubenswrapper[4904]: I0214 11:12:51.062533 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qh6sr" podUID="91b7cb2e-1781-41da-9c90-13b40be0a2b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 14 11:12:52 crc kubenswrapper[4904]: I0214 11:12:52.060138 4904 patch_prober.go:28] interesting pod/router-default-5444994796-qh6sr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 14 11:12:52 crc kubenswrapper[4904]: [-]has-synced failed: reason withheld Feb 14 11:12:52 crc kubenswrapper[4904]: [+]process-running ok Feb 14 11:12:52 crc kubenswrapper[4904]: healthz check failed Feb 14 11:12:52 crc kubenswrapper[4904]: I0214 11:12:52.060515 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qh6sr" podUID="91b7cb2e-1781-41da-9c90-13b40be0a2b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 14 11:12:52 crc kubenswrapper[4904]: I0214 11:12:52.732183 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs\") pod \"network-metrics-daemon-fz5f2\" (UID: \"ac91dd23-e4e5-4073-af76-d760dfdd1adc\") " pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:12:52 crc kubenswrapper[4904]: I0214 11:12:52.738080 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ac91dd23-e4e5-4073-af76-d760dfdd1adc-metrics-certs\") pod \"network-metrics-daemon-fz5f2\" (UID: \"ac91dd23-e4e5-4073-af76-d760dfdd1adc\") " pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:12:52 crc kubenswrapper[4904]: I0214 11:12:52.852145 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fz5f2" Feb 14 11:12:53 crc kubenswrapper[4904]: I0214 11:12:53.058661 4904 patch_prober.go:28] interesting pod/router-default-5444994796-qh6sr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 14 11:12:53 crc kubenswrapper[4904]: [-]has-synced failed: reason withheld Feb 14 11:12:53 crc kubenswrapper[4904]: [+]process-running ok Feb 14 11:12:53 crc kubenswrapper[4904]: healthz check failed Feb 14 11:12:53 crc kubenswrapper[4904]: I0214 11:12:53.058711 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qh6sr" podUID="91b7cb2e-1781-41da-9c90-13b40be0a2b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 14 11:12:54 crc kubenswrapper[4904]: I0214 11:12:54.059798 4904 patch_prober.go:28] interesting pod/router-default-5444994796-qh6sr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 14 11:12:54 crc kubenswrapper[4904]: [-]has-synced failed: reason withheld Feb 14 11:12:54 crc kubenswrapper[4904]: [+]process-running ok Feb 14 11:12:54 crc kubenswrapper[4904]: healthz check failed Feb 14 11:12:54 crc kubenswrapper[4904]: I0214 11:12:54.059929 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qh6sr" podUID="91b7cb2e-1781-41da-9c90-13b40be0a2b4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 14 11:12:55 crc kubenswrapper[4904]: I0214 11:12:55.060173 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:55 crc kubenswrapper[4904]: I0214 11:12:55.068793 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-qh6sr" Feb 14 11:12:59 crc kubenswrapper[4904]: I0214 11:12:59.931273 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-bzps5" Feb 14 11:13:00 crc kubenswrapper[4904]: I0214 11:13:00.864647 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:13:00 crc kubenswrapper[4904]: I0214 11:13:00.868649 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:13:01 crc kubenswrapper[4904]: I0214 11:13:01.586957 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:13:01 crc kubenswrapper[4904]: I0214 11:13:01.928854 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 14 11:13:01 crc kubenswrapper[4904]: I0214 11:13:01.973716 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/617704a5-df59-410d-b6a8-b462453ff137-kubelet-dir\") pod \"617704a5-df59-410d-b6a8-b462453ff137\" (UID: \"617704a5-df59-410d-b6a8-b462453ff137\") " Feb 14 11:13:01 crc kubenswrapper[4904]: I0214 11:13:01.973813 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/617704a5-df59-410d-b6a8-b462453ff137-kube-api-access\") pod \"617704a5-df59-410d-b6a8-b462453ff137\" (UID: \"617704a5-df59-410d-b6a8-b462453ff137\") " Feb 14 11:13:01 crc kubenswrapper[4904]: I0214 11:13:01.973893 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/617704a5-df59-410d-b6a8-b462453ff137-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "617704a5-df59-410d-b6a8-b462453ff137" (UID: "617704a5-df59-410d-b6a8-b462453ff137"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:13:01 crc kubenswrapper[4904]: I0214 11:13:01.974141 4904 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/617704a5-df59-410d-b6a8-b462453ff137-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:01 crc kubenswrapper[4904]: I0214 11:13:01.989086 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/617704a5-df59-410d-b6a8-b462453ff137-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "617704a5-df59-410d-b6a8-b462453ff137" (UID: "617704a5-df59-410d-b6a8-b462453ff137"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:13:02 crc kubenswrapper[4904]: I0214 11:13:02.052772 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"617704a5-df59-410d-b6a8-b462453ff137","Type":"ContainerDied","Data":"7bea0c851dfc564ddf532712a8fe82264af4c70c9a3740b8174c5f28437427c5"} Feb 14 11:13:02 crc kubenswrapper[4904]: I0214 11:13:02.052811 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bea0c851dfc564ddf532712a8fe82264af4c70c9a3740b8174c5f28437427c5" Feb 14 11:13:02 crc kubenswrapper[4904]: I0214 11:13:02.052899 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 14 11:13:02 crc kubenswrapper[4904]: I0214 11:13:02.075447 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/617704a5-df59-410d-b6a8-b462453ff137-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:02 crc kubenswrapper[4904]: I0214 11:13:02.281425 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-fz5f2"] Feb 14 11:13:03 crc kubenswrapper[4904]: I0214 11:13:03.057875 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" event={"ID":"ac91dd23-e4e5-4073-af76-d760dfdd1adc","Type":"ContainerStarted","Data":"b369645c0ac2fb8e6760644278bece2bed0fce4a0c5bd0c93e2db927f82bf5a8"} Feb 14 11:13:04 crc kubenswrapper[4904]: I0214 11:13:04.065913 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" event={"ID":"ac91dd23-e4e5-4073-af76-d760dfdd1adc","Type":"ContainerStarted","Data":"2d50e572186c00279a5633b4019e35d5874b227e999399649b29c482b00411eb"} Feb 14 11:13:10 crc kubenswrapper[4904]: I0214 11:13:10.079087 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fqtd8" Feb 14 11:13:12 crc kubenswrapper[4904]: I0214 11:13:12.093012 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bfn42"] Feb 14 11:13:16 crc kubenswrapper[4904]: I0214 11:13:16.382790 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:13:16 crc kubenswrapper[4904]: I0214 11:13:16.383201 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:13:17 crc kubenswrapper[4904]: E0214 11:13:17.392999 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 14 11:13:17 crc kubenswrapper[4904]: E0214 11:13:17.393369 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h5x2x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-wlsqb_openshift-marketplace(28f09505-e5c3-4105-88d5-9a6d5ed033e9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 14 11:13:17 crc kubenswrapper[4904]: E0214 11:13:17.394563 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-wlsqb" podUID="28f09505-e5c3-4105-88d5-9a6d5ed033e9" Feb 14 11:13:18 crc kubenswrapper[4904]: I0214 11:13:18.267502 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 14 11:13:18 crc kubenswrapper[4904]: E0214 11:13:18.966824 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-wlsqb" podUID="28f09505-e5c3-4105-88d5-9a6d5ed033e9" Feb 14 11:13:19 crc kubenswrapper[4904]: E0214 11:13:19.481891 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 14 11:13:19 crc kubenswrapper[4904]: E0214 11:13:19.482073 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2kf99,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-br8st_openshift-marketplace(f5692d85-aa82-486b-889e-a8964e50b538): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 14 11:13:19 crc kubenswrapper[4904]: E0214 11:13:19.483257 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-br8st" podUID="f5692d85-aa82-486b-889e-a8964e50b538" Feb 14 11:13:20 crc kubenswrapper[4904]: E0214 11:13:20.265632 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-br8st" podUID="f5692d85-aa82-486b-889e-a8964e50b538" Feb 14 11:13:20 crc kubenswrapper[4904]: E0214 11:13:20.505866 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 14 11:13:20 crc kubenswrapper[4904]: E0214 11:13:20.506235 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vw586,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-977kk_openshift-marketplace(92dc9e32-c069-4de6-a3ae-28b80b60d722): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 14 11:13:20 crc kubenswrapper[4904]: E0214 11:13:20.507517 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-977kk" podUID="92dc9e32-c069-4de6-a3ae-28b80b60d722" Feb 14 11:13:21 crc kubenswrapper[4904]: I0214 11:13:21.159736 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nvpkp" event={"ID":"94b8c7dc-4ce9-4434-a387-e19139fe8a3d","Type":"ContainerStarted","Data":"41e52ddb95a6e2f4d6c0f8a62704a328ef23392029f9c947a665d03972e53e12"} Feb 14 11:13:21 crc kubenswrapper[4904]: E0214 11:13:21.161150 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-977kk" podUID="92dc9e32-c069-4de6-a3ae-28b80b60d722" Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.045738 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 14 11:13:22 crc kubenswrapper[4904]: E0214 11:13:22.052696 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="617704a5-df59-410d-b6a8-b462453ff137" containerName="pruner" Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.052813 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="617704a5-df59-410d-b6a8-b462453ff137" containerName="pruner" Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.053063 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="617704a5-df59-410d-b6a8-b462453ff137" containerName="pruner" Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.053599 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.055865 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.056861 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.057314 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5cb2605a-82b5-4dbe-b82c-9885792c664e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5cb2605a-82b5-4dbe-b82c-9885792c664e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.057364 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5cb2605a-82b5-4dbe-b82c-9885792c664e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5cb2605a-82b5-4dbe-b82c-9885792c664e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.063044 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.158087 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5cb2605a-82b5-4dbe-b82c-9885792c664e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5cb2605a-82b5-4dbe-b82c-9885792c664e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.158141 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5cb2605a-82b5-4dbe-b82c-9885792c664e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5cb2605a-82b5-4dbe-b82c-9885792c664e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.158239 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5cb2605a-82b5-4dbe-b82c-9885792c664e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5cb2605a-82b5-4dbe-b82c-9885792c664e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.177982 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5cb2605a-82b5-4dbe-b82c-9885792c664e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5cb2605a-82b5-4dbe-b82c-9885792c664e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.180314 4904 generic.go:334] "Generic (PLEG): container finished" podID="94b8c7dc-4ce9-4434-a387-e19139fe8a3d" containerID="41e52ddb95a6e2f4d6c0f8a62704a328ef23392029f9c947a665d03972e53e12" exitCode=0 Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.180361 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nvpkp" event={"ID":"94b8c7dc-4ce9-4434-a387-e19139fe8a3d","Type":"ContainerDied","Data":"41e52ddb95a6e2f4d6c0f8a62704a328ef23392029f9c947a665d03972e53e12"} Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.185112 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krh2c" event={"ID":"49e91378-b5ff-4480-abbf-765ccc061978","Type":"ContainerStarted","Data":"426f9681e2ee030a5145f3d68a2cbc69bb9e984aaa9607ede33fb69eb9d01750"} Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.186127 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-fz5f2" event={"ID":"ac91dd23-e4e5-4073-af76-d760dfdd1adc","Type":"ContainerStarted","Data":"0e1f1a41de9efd1f385da4d24aac6f50cec2dfa1171d3b0867e096ca9215f77a"} Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.187111 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx9lt" event={"ID":"ee507b1a-9072-40b2-8450-c57209c10a39","Type":"ContainerStarted","Data":"e5ebcf758096f239111708640e2fdd07ccd63e98c2b96c0b7c6abc9660729675"} Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.189280 4904 generic.go:334] "Generic (PLEG): container finished" podID="63c9ce1a-01e3-42f3-8027-bfda9be0b3ff" containerID="566e1b4c9fb87df37d5575f8681b1a8c1074cae72fcf186958f564348cce1b9f" exitCode=0 Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.189304 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bvpm5" event={"ID":"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff","Type":"ContainerDied","Data":"566e1b4c9fb87df37d5575f8681b1a8c1074cae72fcf186958f564348cce1b9f"} Feb 14 11:13:22 crc kubenswrapper[4904]: E0214 11:13:22.310657 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 14 11:13:22 crc kubenswrapper[4904]: E0214 11:13:22.310789 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8jcv6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-xphhn_openshift-marketplace(e3c0e6e8-ac12-4225-b17f-095591f8d2a7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 14 11:13:22 crc kubenswrapper[4904]: E0214 11:13:22.311947 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-xphhn" podUID="e3c0e6e8-ac12-4225-b17f-095591f8d2a7" Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.374132 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 14 11:13:22 crc kubenswrapper[4904]: I0214 11:13:22.764851 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 14 11:13:22 crc kubenswrapper[4904]: W0214 11:13:22.781008 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod5cb2605a_82b5_4dbe_b82c_9885792c664e.slice/crio-cb6a236548c85215551713b3299a7cffd7d8272c9a8065159ee0996aec15dfc8 WatchSource:0}: Error finding container cb6a236548c85215551713b3299a7cffd7d8272c9a8065159ee0996aec15dfc8: Status 404 returned error can't find the container with id cb6a236548c85215551713b3299a7cffd7d8272c9a8065159ee0996aec15dfc8 Feb 14 11:13:23 crc kubenswrapper[4904]: I0214 11:13:23.207277 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"5cb2605a-82b5-4dbe-b82c-9885792c664e","Type":"ContainerStarted","Data":"168298ffbcce922b43c998655592995eab9f534e24c9d2755e683dd4fb14c4dc"} Feb 14 11:13:23 crc kubenswrapper[4904]: I0214 11:13:23.207570 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"5cb2605a-82b5-4dbe-b82c-9885792c664e","Type":"ContainerStarted","Data":"cb6a236548c85215551713b3299a7cffd7d8272c9a8065159ee0996aec15dfc8"} Feb 14 11:13:23 crc kubenswrapper[4904]: I0214 11:13:23.209034 4904 generic.go:334] "Generic (PLEG): container finished" podID="49e91378-b5ff-4480-abbf-765ccc061978" containerID="426f9681e2ee030a5145f3d68a2cbc69bb9e984aaa9607ede33fb69eb9d01750" exitCode=0 Feb 14 11:13:23 crc kubenswrapper[4904]: I0214 11:13:23.209883 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krh2c" event={"ID":"49e91378-b5ff-4480-abbf-765ccc061978","Type":"ContainerDied","Data":"426f9681e2ee030a5145f3d68a2cbc69bb9e984aaa9607ede33fb69eb9d01750"} Feb 14 11:13:23 crc kubenswrapper[4904]: I0214 11:13:23.213845 4904 generic.go:334] "Generic (PLEG): container finished" podID="ee507b1a-9072-40b2-8450-c57209c10a39" containerID="e5ebcf758096f239111708640e2fdd07ccd63e98c2b96c0b7c6abc9660729675" exitCode=0 Feb 14 11:13:23 crc kubenswrapper[4904]: I0214 11:13:23.214070 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx9lt" event={"ID":"ee507b1a-9072-40b2-8450-c57209c10a39","Type":"ContainerDied","Data":"e5ebcf758096f239111708640e2fdd07ccd63e98c2b96c0b7c6abc9660729675"} Feb 14 11:13:23 crc kubenswrapper[4904]: E0214 11:13:23.219892 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-xphhn" podUID="e3c0e6e8-ac12-4225-b17f-095591f8d2a7" Feb 14 11:13:23 crc kubenswrapper[4904]: I0214 11:13:23.227502 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=1.227487028 podStartE2EDuration="1.227487028s" podCreationTimestamp="2026-02-14 11:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:13:23.220560315 +0000 UTC m=+194.033324976" watchObservedRunningTime="2026-02-14 11:13:23.227487028 +0000 UTC m=+194.040251689" Feb 14 11:13:23 crc kubenswrapper[4904]: I0214 11:13:23.262717 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-fz5f2" podStartSLOduration=173.262699148 podStartE2EDuration="2m53.262699148s" podCreationTimestamp="2026-02-14 11:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:13:23.258496457 +0000 UTC m=+194.071261118" watchObservedRunningTime="2026-02-14 11:13:23.262699148 +0000 UTC m=+194.075463809" Feb 14 11:13:24 crc kubenswrapper[4904]: I0214 11:13:24.221135 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nvpkp" event={"ID":"94b8c7dc-4ce9-4434-a387-e19139fe8a3d","Type":"ContainerStarted","Data":"91278d384c32444678079c8fda4688f16a28bc0032d085b4a9216e1b1801b357"} Feb 14 11:13:24 crc kubenswrapper[4904]: I0214 11:13:24.222977 4904 generic.go:334] "Generic (PLEG): container finished" podID="5cb2605a-82b5-4dbe-b82c-9885792c664e" containerID="168298ffbcce922b43c998655592995eab9f534e24c9d2755e683dd4fb14c4dc" exitCode=0 Feb 14 11:13:24 crc kubenswrapper[4904]: I0214 11:13:24.223084 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"5cb2605a-82b5-4dbe-b82c-9885792c664e","Type":"ContainerDied","Data":"168298ffbcce922b43c998655592995eab9f534e24c9d2755e683dd4fb14c4dc"} Feb 14 11:13:24 crc kubenswrapper[4904]: I0214 11:13:24.225723 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krh2c" event={"ID":"49e91378-b5ff-4480-abbf-765ccc061978","Type":"ContainerStarted","Data":"57d5b7a6e0f85a669f18f54f74eccce27d403d2021bf46561a6cacc13cc5beb6"} Feb 14 11:13:24 crc kubenswrapper[4904]: I0214 11:13:24.228881 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx9lt" event={"ID":"ee507b1a-9072-40b2-8450-c57209c10a39","Type":"ContainerStarted","Data":"a66dc168e7a9a5e934b79fb11756561a88e9a2fb3539909d091c66f87ac433c6"} Feb 14 11:13:24 crc kubenswrapper[4904]: I0214 11:13:24.230969 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bvpm5" event={"ID":"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff","Type":"ContainerStarted","Data":"6c243d2633d6cd00022d141ab70d97df2920850cf29e1c198bbc7684c36ac17f"} Feb 14 11:13:24 crc kubenswrapper[4904]: I0214 11:13:24.245205 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nvpkp" podStartSLOduration=3.8470700190000002 podStartE2EDuration="45.245187504s" podCreationTimestamp="2026-02-14 11:12:39 +0000 UTC" firstStartedPulling="2026-02-14 11:12:41.566663395 +0000 UTC m=+152.379428046" lastFinishedPulling="2026-02-14 11:13:22.96478085 +0000 UTC m=+193.777545531" observedRunningTime="2026-02-14 11:13:24.241374023 +0000 UTC m=+195.054138684" watchObservedRunningTime="2026-02-14 11:13:24.245187504 +0000 UTC m=+195.057952165" Feb 14 11:13:24 crc kubenswrapper[4904]: I0214 11:13:24.269545 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zx9lt" podStartSLOduration=3.119617239 podStartE2EDuration="46.269517526s" podCreationTimestamp="2026-02-14 11:12:38 +0000 UTC" firstStartedPulling="2026-02-14 11:12:40.468576556 +0000 UTC m=+151.281341217" lastFinishedPulling="2026-02-14 11:13:23.618476843 +0000 UTC m=+194.431241504" observedRunningTime="2026-02-14 11:13:24.264920915 +0000 UTC m=+195.077685596" watchObservedRunningTime="2026-02-14 11:13:24.269517526 +0000 UTC m=+195.082282187" Feb 14 11:13:24 crc kubenswrapper[4904]: I0214 11:13:24.322590 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-krh2c" podStartSLOduration=3.26968101 podStartE2EDuration="44.322561597s" podCreationTimestamp="2026-02-14 11:12:40 +0000 UTC" firstStartedPulling="2026-02-14 11:12:42.6339086 +0000 UTC m=+153.446673261" lastFinishedPulling="2026-02-14 11:13:23.686789187 +0000 UTC m=+194.499553848" observedRunningTime="2026-02-14 11:13:24.321036647 +0000 UTC m=+195.133801308" watchObservedRunningTime="2026-02-14 11:13:24.322561597 +0000 UTC m=+195.135326258" Feb 14 11:13:24 crc kubenswrapper[4904]: I0214 11:13:24.346083 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bvpm5" podStartSLOduration=4.180502468 podStartE2EDuration="47.346054668s" podCreationTimestamp="2026-02-14 11:12:37 +0000 UTC" firstStartedPulling="2026-02-14 11:12:40.486298235 +0000 UTC m=+151.299062896" lastFinishedPulling="2026-02-14 11:13:23.651850435 +0000 UTC m=+194.464615096" observedRunningTime="2026-02-14 11:13:24.34350223 +0000 UTC m=+195.156266891" watchObservedRunningTime="2026-02-14 11:13:24.346054668 +0000 UTC m=+195.158819329" Feb 14 11:13:25 crc kubenswrapper[4904]: I0214 11:13:25.754171 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 14 11:13:25 crc kubenswrapper[4904]: I0214 11:13:25.912240 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5cb2605a-82b5-4dbe-b82c-9885792c664e-kube-api-access\") pod \"5cb2605a-82b5-4dbe-b82c-9885792c664e\" (UID: \"5cb2605a-82b5-4dbe-b82c-9885792c664e\") " Feb 14 11:13:25 crc kubenswrapper[4904]: I0214 11:13:25.912401 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5cb2605a-82b5-4dbe-b82c-9885792c664e-kubelet-dir\") pod \"5cb2605a-82b5-4dbe-b82c-9885792c664e\" (UID: \"5cb2605a-82b5-4dbe-b82c-9885792c664e\") " Feb 14 11:13:25 crc kubenswrapper[4904]: I0214 11:13:25.912558 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5cb2605a-82b5-4dbe-b82c-9885792c664e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5cb2605a-82b5-4dbe-b82c-9885792c664e" (UID: "5cb2605a-82b5-4dbe-b82c-9885792c664e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:13:25 crc kubenswrapper[4904]: I0214 11:13:25.922723 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cb2605a-82b5-4dbe-b82c-9885792c664e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5cb2605a-82b5-4dbe-b82c-9885792c664e" (UID: "5cb2605a-82b5-4dbe-b82c-9885792c664e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:13:26 crc kubenswrapper[4904]: I0214 11:13:26.013859 4904 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5cb2605a-82b5-4dbe-b82c-9885792c664e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:26 crc kubenswrapper[4904]: I0214 11:13:26.013900 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5cb2605a-82b5-4dbe-b82c-9885792c664e-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:26 crc kubenswrapper[4904]: I0214 11:13:26.241827 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"5cb2605a-82b5-4dbe-b82c-9885792c664e","Type":"ContainerDied","Data":"cb6a236548c85215551713b3299a7cffd7d8272c9a8065159ee0996aec15dfc8"} Feb 14 11:13:26 crc kubenswrapper[4904]: I0214 11:13:26.241874 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb6a236548c85215551713b3299a7cffd7d8272c9a8065159ee0996aec15dfc8" Feb 14 11:13:26 crc kubenswrapper[4904]: I0214 11:13:26.241969 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 14 11:13:28 crc kubenswrapper[4904]: I0214 11:13:28.223299 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bvpm5" Feb 14 11:13:28 crc kubenswrapper[4904]: I0214 11:13:28.223363 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bvpm5" Feb 14 11:13:28 crc kubenswrapper[4904]: I0214 11:13:28.501243 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bvpm5" Feb 14 11:13:28 crc kubenswrapper[4904]: I0214 11:13:28.539146 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zx9lt" Feb 14 11:13:28 crc kubenswrapper[4904]: I0214 11:13:28.539397 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zx9lt" Feb 14 11:13:28 crc kubenswrapper[4904]: I0214 11:13:28.551113 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bvpm5" Feb 14 11:13:28 crc kubenswrapper[4904]: I0214 11:13:28.601785 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zx9lt" Feb 14 11:13:29 crc kubenswrapper[4904]: I0214 11:13:29.292799 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zx9lt" Feb 14 11:13:29 crc kubenswrapper[4904]: I0214 11:13:29.444062 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 14 11:13:29 crc kubenswrapper[4904]: E0214 11:13:29.444412 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb2605a-82b5-4dbe-b82c-9885792c664e" containerName="pruner" Feb 14 11:13:29 crc kubenswrapper[4904]: I0214 11:13:29.444427 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb2605a-82b5-4dbe-b82c-9885792c664e" containerName="pruner" Feb 14 11:13:29 crc kubenswrapper[4904]: I0214 11:13:29.444615 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cb2605a-82b5-4dbe-b82c-9885792c664e" containerName="pruner" Feb 14 11:13:29 crc kubenswrapper[4904]: I0214 11:13:29.445116 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 14 11:13:29 crc kubenswrapper[4904]: I0214 11:13:29.452126 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 14 11:13:29 crc kubenswrapper[4904]: I0214 11:13:29.452450 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 14 11:13:29 crc kubenswrapper[4904]: I0214 11:13:29.463870 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 14 11:13:29 crc kubenswrapper[4904]: I0214 11:13:29.464325 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 14 11:13:29 crc kubenswrapper[4904]: I0214 11:13:29.464382 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96-kube-api-access\") pod \"installer-9-crc\" (UID: \"b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 14 11:13:29 crc kubenswrapper[4904]: I0214 11:13:29.464597 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96-var-lock\") pod \"installer-9-crc\" (UID: \"b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 14 11:13:29 crc kubenswrapper[4904]: I0214 11:13:29.565654 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96-var-lock\") pod \"installer-9-crc\" (UID: \"b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 14 11:13:29 crc kubenswrapper[4904]: I0214 11:13:29.565714 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 14 11:13:29 crc kubenswrapper[4904]: I0214 11:13:29.565754 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96-kube-api-access\") pod \"installer-9-crc\" (UID: \"b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 14 11:13:29 crc kubenswrapper[4904]: I0214 11:13:29.565796 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96-var-lock\") pod \"installer-9-crc\" (UID: \"b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 14 11:13:29 crc kubenswrapper[4904]: I0214 11:13:29.565796 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 14 11:13:29 crc kubenswrapper[4904]: I0214 11:13:29.587578 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96-kube-api-access\") pod \"installer-9-crc\" (UID: \"b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 14 11:13:29 crc kubenswrapper[4904]: I0214 11:13:29.775302 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 14 11:13:30 crc kubenswrapper[4904]: I0214 11:13:30.192628 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 14 11:13:30 crc kubenswrapper[4904]: I0214 11:13:30.264470 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96","Type":"ContainerStarted","Data":"c886e817586198f4cdc789e363ad7d5558146618ebce84ef0fc439c9b47902c9"} Feb 14 11:13:30 crc kubenswrapper[4904]: I0214 11:13:30.327207 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nvpkp" Feb 14 11:13:30 crc kubenswrapper[4904]: I0214 11:13:30.327272 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nvpkp" Feb 14 11:13:30 crc kubenswrapper[4904]: I0214 11:13:30.371530 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nvpkp" Feb 14 11:13:30 crc kubenswrapper[4904]: I0214 11:13:30.451481 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zx9lt"] Feb 14 11:13:31 crc kubenswrapper[4904]: I0214 11:13:31.271943 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96","Type":"ContainerStarted","Data":"1b9061fcc31f6d7899b85f1e9b619bb1363000d702e443f75da43fab2b7fb08e"} Feb 14 11:13:31 crc kubenswrapper[4904]: I0214 11:13:31.272502 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zx9lt" podUID="ee507b1a-9072-40b2-8450-c57209c10a39" containerName="registry-server" containerID="cri-o://a66dc168e7a9a5e934b79fb11756561a88e9a2fb3539909d091c66f87ac433c6" gracePeriod=2 Feb 14 11:13:31 crc kubenswrapper[4904]: I0214 11:13:31.308124 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.308106042 podStartE2EDuration="2.308106042s" podCreationTimestamp="2026-02-14 11:13:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:13:31.295485308 +0000 UTC m=+202.108249969" watchObservedRunningTime="2026-02-14 11:13:31.308106042 +0000 UTC m=+202.120870703" Feb 14 11:13:31 crc kubenswrapper[4904]: I0214 11:13:31.330561 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nvpkp" Feb 14 11:13:31 crc kubenswrapper[4904]: I0214 11:13:31.395812 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-krh2c" Feb 14 11:13:31 crc kubenswrapper[4904]: I0214 11:13:31.396195 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-krh2c" Feb 14 11:13:31 crc kubenswrapper[4904]: I0214 11:13:31.446140 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-krh2c" Feb 14 11:13:31 crc kubenswrapper[4904]: I0214 11:13:31.712246 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx9lt" Feb 14 11:13:31 crc kubenswrapper[4904]: I0214 11:13:31.892761 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjxn6\" (UniqueName: \"kubernetes.io/projected/ee507b1a-9072-40b2-8450-c57209c10a39-kube-api-access-qjxn6\") pod \"ee507b1a-9072-40b2-8450-c57209c10a39\" (UID: \"ee507b1a-9072-40b2-8450-c57209c10a39\") " Feb 14 11:13:31 crc kubenswrapper[4904]: I0214 11:13:31.893263 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee507b1a-9072-40b2-8450-c57209c10a39-utilities\") pod \"ee507b1a-9072-40b2-8450-c57209c10a39\" (UID: \"ee507b1a-9072-40b2-8450-c57209c10a39\") " Feb 14 11:13:31 crc kubenswrapper[4904]: I0214 11:13:31.893400 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee507b1a-9072-40b2-8450-c57209c10a39-catalog-content\") pod \"ee507b1a-9072-40b2-8450-c57209c10a39\" (UID: \"ee507b1a-9072-40b2-8450-c57209c10a39\") " Feb 14 11:13:31 crc kubenswrapper[4904]: I0214 11:13:31.894528 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee507b1a-9072-40b2-8450-c57209c10a39-utilities" (OuterVolumeSpecName: "utilities") pod "ee507b1a-9072-40b2-8450-c57209c10a39" (UID: "ee507b1a-9072-40b2-8450-c57209c10a39"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:13:31 crc kubenswrapper[4904]: I0214 11:13:31.905032 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee507b1a-9072-40b2-8450-c57209c10a39-kube-api-access-qjxn6" (OuterVolumeSpecName: "kube-api-access-qjxn6") pod "ee507b1a-9072-40b2-8450-c57209c10a39" (UID: "ee507b1a-9072-40b2-8450-c57209c10a39"). InnerVolumeSpecName "kube-api-access-qjxn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:13:31 crc kubenswrapper[4904]: I0214 11:13:31.957407 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee507b1a-9072-40b2-8450-c57209c10a39-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee507b1a-9072-40b2-8450-c57209c10a39" (UID: "ee507b1a-9072-40b2-8450-c57209c10a39"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:13:31 crc kubenswrapper[4904]: I0214 11:13:31.994929 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee507b1a-9072-40b2-8450-c57209c10a39-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:31 crc kubenswrapper[4904]: I0214 11:13:31.994971 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjxn6\" (UniqueName: \"kubernetes.io/projected/ee507b1a-9072-40b2-8450-c57209c10a39-kube-api-access-qjxn6\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:31 crc kubenswrapper[4904]: I0214 11:13:31.994987 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee507b1a-9072-40b2-8450-c57209c10a39-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:32 crc kubenswrapper[4904]: I0214 11:13:32.279939 4904 generic.go:334] "Generic (PLEG): container finished" podID="28f09505-e5c3-4105-88d5-9a6d5ed033e9" containerID="e4ffd53226b042342190efa7d0897a5537d1ac98dca6af334e95d851e7c4e81f" exitCode=0 Feb 14 11:13:32 crc kubenswrapper[4904]: I0214 11:13:32.280376 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlsqb" event={"ID":"28f09505-e5c3-4105-88d5-9a6d5ed033e9","Type":"ContainerDied","Data":"e4ffd53226b042342190efa7d0897a5537d1ac98dca6af334e95d851e7c4e81f"} Feb 14 11:13:32 crc kubenswrapper[4904]: I0214 11:13:32.282857 4904 generic.go:334] "Generic (PLEG): container finished" podID="ee507b1a-9072-40b2-8450-c57209c10a39" containerID="a66dc168e7a9a5e934b79fb11756561a88e9a2fb3539909d091c66f87ac433c6" exitCode=0 Feb 14 11:13:32 crc kubenswrapper[4904]: I0214 11:13:32.282910 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx9lt" Feb 14 11:13:32 crc kubenswrapper[4904]: I0214 11:13:32.282923 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx9lt" event={"ID":"ee507b1a-9072-40b2-8450-c57209c10a39","Type":"ContainerDied","Data":"a66dc168e7a9a5e934b79fb11756561a88e9a2fb3539909d091c66f87ac433c6"} Feb 14 11:13:32 crc kubenswrapper[4904]: I0214 11:13:32.283110 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx9lt" event={"ID":"ee507b1a-9072-40b2-8450-c57209c10a39","Type":"ContainerDied","Data":"e5a24a82a2b61dd4adef5d9a31562decd67cbd2548dd8575d54907cddd642297"} Feb 14 11:13:32 crc kubenswrapper[4904]: I0214 11:13:32.283132 4904 scope.go:117] "RemoveContainer" containerID="a66dc168e7a9a5e934b79fb11756561a88e9a2fb3539909d091c66f87ac433c6" Feb 14 11:13:32 crc kubenswrapper[4904]: I0214 11:13:32.306441 4904 scope.go:117] "RemoveContainer" containerID="e5ebcf758096f239111708640e2fdd07ccd63e98c2b96c0b7c6abc9660729675" Feb 14 11:13:32 crc kubenswrapper[4904]: I0214 11:13:32.331118 4904 scope.go:117] "RemoveContainer" containerID="2273b2c80df3f5550f5b4463f68201663f6da1d8dff7a85cd8258b45fb817db7" Feb 14 11:13:32 crc kubenswrapper[4904]: I0214 11:13:32.336918 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zx9lt"] Feb 14 11:13:32 crc kubenswrapper[4904]: I0214 11:13:32.340524 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zx9lt"] Feb 14 11:13:32 crc kubenswrapper[4904]: I0214 11:13:32.343633 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-krh2c" Feb 14 11:13:32 crc kubenswrapper[4904]: I0214 11:13:32.349793 4904 scope.go:117] "RemoveContainer" containerID="a66dc168e7a9a5e934b79fb11756561a88e9a2fb3539909d091c66f87ac433c6" Feb 14 11:13:32 crc kubenswrapper[4904]: E0214 11:13:32.350666 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a66dc168e7a9a5e934b79fb11756561a88e9a2fb3539909d091c66f87ac433c6\": container with ID starting with a66dc168e7a9a5e934b79fb11756561a88e9a2fb3539909d091c66f87ac433c6 not found: ID does not exist" containerID="a66dc168e7a9a5e934b79fb11756561a88e9a2fb3539909d091c66f87ac433c6" Feb 14 11:13:32 crc kubenswrapper[4904]: I0214 11:13:32.350825 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a66dc168e7a9a5e934b79fb11756561a88e9a2fb3539909d091c66f87ac433c6"} err="failed to get container status \"a66dc168e7a9a5e934b79fb11756561a88e9a2fb3539909d091c66f87ac433c6\": rpc error: code = NotFound desc = could not find container \"a66dc168e7a9a5e934b79fb11756561a88e9a2fb3539909d091c66f87ac433c6\": container with ID starting with a66dc168e7a9a5e934b79fb11756561a88e9a2fb3539909d091c66f87ac433c6 not found: ID does not exist" Feb 14 11:13:32 crc kubenswrapper[4904]: I0214 11:13:32.350927 4904 scope.go:117] "RemoveContainer" containerID="e5ebcf758096f239111708640e2fdd07ccd63e98c2b96c0b7c6abc9660729675" Feb 14 11:13:32 crc kubenswrapper[4904]: E0214 11:13:32.351698 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5ebcf758096f239111708640e2fdd07ccd63e98c2b96c0b7c6abc9660729675\": container with ID starting with e5ebcf758096f239111708640e2fdd07ccd63e98c2b96c0b7c6abc9660729675 not found: ID does not exist" containerID="e5ebcf758096f239111708640e2fdd07ccd63e98c2b96c0b7c6abc9660729675" Feb 14 11:13:32 crc kubenswrapper[4904]: I0214 11:13:32.351726 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5ebcf758096f239111708640e2fdd07ccd63e98c2b96c0b7c6abc9660729675"} err="failed to get container status \"e5ebcf758096f239111708640e2fdd07ccd63e98c2b96c0b7c6abc9660729675\": rpc error: code = NotFound desc = could not find container \"e5ebcf758096f239111708640e2fdd07ccd63e98c2b96c0b7c6abc9660729675\": container with ID starting with e5ebcf758096f239111708640e2fdd07ccd63e98c2b96c0b7c6abc9660729675 not found: ID does not exist" Feb 14 11:13:32 crc kubenswrapper[4904]: I0214 11:13:32.351747 4904 scope.go:117] "RemoveContainer" containerID="2273b2c80df3f5550f5b4463f68201663f6da1d8dff7a85cd8258b45fb817db7" Feb 14 11:13:32 crc kubenswrapper[4904]: E0214 11:13:32.352011 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2273b2c80df3f5550f5b4463f68201663f6da1d8dff7a85cd8258b45fb817db7\": container with ID starting with 2273b2c80df3f5550f5b4463f68201663f6da1d8dff7a85cd8258b45fb817db7 not found: ID does not exist" containerID="2273b2c80df3f5550f5b4463f68201663f6da1d8dff7a85cd8258b45fb817db7" Feb 14 11:13:32 crc kubenswrapper[4904]: I0214 11:13:32.352039 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2273b2c80df3f5550f5b4463f68201663f6da1d8dff7a85cd8258b45fb817db7"} err="failed to get container status \"2273b2c80df3f5550f5b4463f68201663f6da1d8dff7a85cd8258b45fb817db7\": rpc error: code = NotFound desc = could not find container \"2273b2c80df3f5550f5b4463f68201663f6da1d8dff7a85cd8258b45fb817db7\": container with ID starting with 2273b2c80df3f5550f5b4463f68201663f6da1d8dff7a85cd8258b45fb817db7 not found: ID does not exist" Feb 14 11:13:33 crc kubenswrapper[4904]: I0214 11:13:33.292890 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlsqb" event={"ID":"28f09505-e5c3-4105-88d5-9a6d5ed033e9","Type":"ContainerStarted","Data":"778442d1f9bdc3c507a90be7b2083d3b93f728b5a9e12f99fef04a8a5601fc6a"} Feb 14 11:13:33 crc kubenswrapper[4904]: I0214 11:13:33.847482 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee507b1a-9072-40b2-8450-c57209c10a39" path="/var/lib/kubelet/pods/ee507b1a-9072-40b2-8450-c57209c10a39/volumes" Feb 14 11:13:33 crc kubenswrapper[4904]: I0214 11:13:33.861513 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wlsqb" podStartSLOduration=4.740431582 podStartE2EDuration="55.861491544s" podCreationTimestamp="2026-02-14 11:12:38 +0000 UTC" firstStartedPulling="2026-02-14 11:12:41.565039962 +0000 UTC m=+152.377804623" lastFinishedPulling="2026-02-14 11:13:32.686099884 +0000 UTC m=+203.498864585" observedRunningTime="2026-02-14 11:13:33.31687585 +0000 UTC m=+204.129640521" watchObservedRunningTime="2026-02-14 11:13:33.861491544 +0000 UTC m=+204.674256205" Feb 14 11:13:34 crc kubenswrapper[4904]: I0214 11:13:34.302183 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-br8st" event={"ID":"f5692d85-aa82-486b-889e-a8964e50b538","Type":"ContainerStarted","Data":"b246c75b12d3a74095e39729b2a13c83b79500ec4e92b1f9ddda3a5ab73ba2df"} Feb 14 11:13:35 crc kubenswrapper[4904]: I0214 11:13:35.311663 4904 generic.go:334] "Generic (PLEG): container finished" podID="f5692d85-aa82-486b-889e-a8964e50b538" containerID="b246c75b12d3a74095e39729b2a13c83b79500ec4e92b1f9ddda3a5ab73ba2df" exitCode=0 Feb 14 11:13:35 crc kubenswrapper[4904]: I0214 11:13:35.311763 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-br8st" event={"ID":"f5692d85-aa82-486b-889e-a8964e50b538","Type":"ContainerDied","Data":"b246c75b12d3a74095e39729b2a13c83b79500ec4e92b1f9ddda3a5ab73ba2df"} Feb 14 11:13:36 crc kubenswrapper[4904]: I0214 11:13:36.322290 4904 generic.go:334] "Generic (PLEG): container finished" podID="92dc9e32-c069-4de6-a3ae-28b80b60d722" containerID="866702745cde4a085d2378b9245f45e6594a426b09b695b26c815e7de28e2211" exitCode=0 Feb 14 11:13:36 crc kubenswrapper[4904]: I0214 11:13:36.322412 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-977kk" event={"ID":"92dc9e32-c069-4de6-a3ae-28b80b60d722","Type":"ContainerDied","Data":"866702745cde4a085d2378b9245f45e6594a426b09b695b26c815e7de28e2211"} Feb 14 11:13:36 crc kubenswrapper[4904]: I0214 11:13:36.327823 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-br8st" event={"ID":"f5692d85-aa82-486b-889e-a8964e50b538","Type":"ContainerStarted","Data":"9f129083f5d4e701e483dedf6c257b7b435696006cdf5aa9c0dadfe1fa3d9d04"} Feb 14 11:13:36 crc kubenswrapper[4904]: I0214 11:13:36.366336 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-br8st" podStartSLOduration=3.366365974 podStartE2EDuration="56.36630952s" podCreationTimestamp="2026-02-14 11:12:40 +0000 UTC" firstStartedPulling="2026-02-14 11:12:42.691431639 +0000 UTC m=+153.504196300" lastFinishedPulling="2026-02-14 11:13:35.691375175 +0000 UTC m=+206.504139846" observedRunningTime="2026-02-14 11:13:36.364252522 +0000 UTC m=+207.177017203" watchObservedRunningTime="2026-02-14 11:13:36.36630952 +0000 UTC m=+207.179074181" Feb 14 11:13:37 crc kubenswrapper[4904]: I0214 11:13:37.120203 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" podUID="e831e54b-3826-4415-9b43-ce02e5ed81bf" containerName="oauth-openshift" containerID="cri-o://0a42a22bc2927f70424cd130fee97a557db035a17cdc29afff69644a2db61987" gracePeriod=15 Feb 14 11:13:37 crc kubenswrapper[4904]: I0214 11:13:37.334161 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-977kk" event={"ID":"92dc9e32-c069-4de6-a3ae-28b80b60d722","Type":"ContainerStarted","Data":"62d391f394fced8cef03cbba49f457f6818005d1c190e5ca767d45bbc9d57a2c"} Feb 14 11:13:37 crc kubenswrapper[4904]: I0214 11:13:37.335680 4904 generic.go:334] "Generic (PLEG): container finished" podID="e831e54b-3826-4415-9b43-ce02e5ed81bf" containerID="0a42a22bc2927f70424cd130fee97a557db035a17cdc29afff69644a2db61987" exitCode=0 Feb 14 11:13:37 crc kubenswrapper[4904]: I0214 11:13:37.335714 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" event={"ID":"e831e54b-3826-4415-9b43-ce02e5ed81bf","Type":"ContainerDied","Data":"0a42a22bc2927f70424cd130fee97a557db035a17cdc29afff69644a2db61987"} Feb 14 11:13:37 crc kubenswrapper[4904]: I0214 11:13:37.993218 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.067487 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-69b55d54f6-bnplh"] Feb 14 11:13:38 crc kubenswrapper[4904]: E0214 11:13:38.067674 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee507b1a-9072-40b2-8450-c57209c10a39" containerName="extract-utilities" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.067686 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee507b1a-9072-40b2-8450-c57209c10a39" containerName="extract-utilities" Feb 14 11:13:38 crc kubenswrapper[4904]: E0214 11:13:38.067699 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e831e54b-3826-4415-9b43-ce02e5ed81bf" containerName="oauth-openshift" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.067705 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="e831e54b-3826-4415-9b43-ce02e5ed81bf" containerName="oauth-openshift" Feb 14 11:13:38 crc kubenswrapper[4904]: E0214 11:13:38.067713 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee507b1a-9072-40b2-8450-c57209c10a39" containerName="extract-content" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.067719 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee507b1a-9072-40b2-8450-c57209c10a39" containerName="extract-content" Feb 14 11:13:38 crc kubenswrapper[4904]: E0214 11:13:38.067736 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee507b1a-9072-40b2-8450-c57209c10a39" containerName="registry-server" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.067742 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee507b1a-9072-40b2-8450-c57209c10a39" containerName="registry-server" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.067851 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="e831e54b-3826-4415-9b43-ce02e5ed81bf" containerName="oauth-openshift" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.067866 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee507b1a-9072-40b2-8450-c57209c10a39" containerName="registry-server" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.068194 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087255 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-ocp-branding-template\") pod \"e831e54b-3826-4415-9b43-ce02e5ed81bf\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087289 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-idp-0-file-data\") pod \"e831e54b-3826-4415-9b43-ce02e5ed81bf\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087317 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-service-ca\") pod \"e831e54b-3826-4415-9b43-ce02e5ed81bf\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087337 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhb2k\" (UniqueName: \"kubernetes.io/projected/e831e54b-3826-4415-9b43-ce02e5ed81bf-kube-api-access-nhb2k\") pod \"e831e54b-3826-4415-9b43-ce02e5ed81bf\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087359 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-template-provider-selection\") pod \"e831e54b-3826-4415-9b43-ce02e5ed81bf\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087375 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-template-login\") pod \"e831e54b-3826-4415-9b43-ce02e5ed81bf\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087391 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-audit-policies\") pod \"e831e54b-3826-4415-9b43-ce02e5ed81bf\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087412 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-template-error\") pod \"e831e54b-3826-4415-9b43-ce02e5ed81bf\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087429 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-session\") pod \"e831e54b-3826-4415-9b43-ce02e5ed81bf\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087445 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-trusted-ca-bundle\") pod \"e831e54b-3826-4415-9b43-ce02e5ed81bf\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087459 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e831e54b-3826-4415-9b43-ce02e5ed81bf-audit-dir\") pod \"e831e54b-3826-4415-9b43-ce02e5ed81bf\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087473 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-serving-cert\") pod \"e831e54b-3826-4415-9b43-ce02e5ed81bf\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087510 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-router-certs\") pod \"e831e54b-3826-4415-9b43-ce02e5ed81bf\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087536 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-cliconfig\") pod \"e831e54b-3826-4415-9b43-ce02e5ed81bf\" (UID: \"e831e54b-3826-4415-9b43-ce02e5ed81bf\") " Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087643 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-service-ca\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087663 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-serving-cert\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087685 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e8cddb62-7172-45e1-b162-ae9360043f92-audit-policies\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087698 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-session\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087713 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087734 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087753 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e8cddb62-7172-45e1-b162-ae9360043f92-audit-dir\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087768 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087793 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5nm5\" (UniqueName: \"kubernetes.io/projected/e8cddb62-7172-45e1-b162-ae9360043f92-kube-api-access-x5nm5\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087813 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-cliconfig\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087827 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-router-certs\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087865 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087881 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-user-template-error\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.087905 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-user-template-login\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.089019 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "e831e54b-3826-4415-9b43-ce02e5ed81bf" (UID: "e831e54b-3826-4415-9b43-ce02e5ed81bf"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.089096 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "e831e54b-3826-4415-9b43-ce02e5ed81bf" (UID: "e831e54b-3826-4415-9b43-ce02e5ed81bf"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.089150 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e831e54b-3826-4415-9b43-ce02e5ed81bf-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "e831e54b-3826-4415-9b43-ce02e5ed81bf" (UID: "e831e54b-3826-4415-9b43-ce02e5ed81bf"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.089202 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "e831e54b-3826-4415-9b43-ce02e5ed81bf" (UID: "e831e54b-3826-4415-9b43-ce02e5ed81bf"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.089211 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "e831e54b-3826-4415-9b43-ce02e5ed81bf" (UID: "e831e54b-3826-4415-9b43-ce02e5ed81bf"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.108530 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "e831e54b-3826-4415-9b43-ce02e5ed81bf" (UID: "e831e54b-3826-4415-9b43-ce02e5ed81bf"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.121724 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "e831e54b-3826-4415-9b43-ce02e5ed81bf" (UID: "e831e54b-3826-4415-9b43-ce02e5ed81bf"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.130393 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "e831e54b-3826-4415-9b43-ce02e5ed81bf" (UID: "e831e54b-3826-4415-9b43-ce02e5ed81bf"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.131045 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e831e54b-3826-4415-9b43-ce02e5ed81bf-kube-api-access-nhb2k" (OuterVolumeSpecName: "kube-api-access-nhb2k") pod "e831e54b-3826-4415-9b43-ce02e5ed81bf" (UID: "e831e54b-3826-4415-9b43-ce02e5ed81bf"). InnerVolumeSpecName "kube-api-access-nhb2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.131144 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "e831e54b-3826-4415-9b43-ce02e5ed81bf" (UID: "e831e54b-3826-4415-9b43-ce02e5ed81bf"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.134242 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "e831e54b-3826-4415-9b43-ce02e5ed81bf" (UID: "e831e54b-3826-4415-9b43-ce02e5ed81bf"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.140079 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-69b55d54f6-bnplh"] Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.141761 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "e831e54b-3826-4415-9b43-ce02e5ed81bf" (UID: "e831e54b-3826-4415-9b43-ce02e5ed81bf"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.143976 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "e831e54b-3826-4415-9b43-ce02e5ed81bf" (UID: "e831e54b-3826-4415-9b43-ce02e5ed81bf"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.144699 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "e831e54b-3826-4415-9b43-ce02e5ed81bf" (UID: "e831e54b-3826-4415-9b43-ce02e5ed81bf"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.188945 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189002 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-user-template-error\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189033 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-user-template-login\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189076 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-service-ca\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189094 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-serving-cert\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189119 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e8cddb62-7172-45e1-b162-ae9360043f92-audit-policies\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189135 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-session\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189149 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189173 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189194 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e8cddb62-7172-45e1-b162-ae9360043f92-audit-dir\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189209 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189233 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5nm5\" (UniqueName: \"kubernetes.io/projected/e8cddb62-7172-45e1-b162-ae9360043f92-kube-api-access-x5nm5\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189254 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-cliconfig\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189269 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-router-certs\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189305 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhb2k\" (UniqueName: \"kubernetes.io/projected/e831e54b-3826-4415-9b43-ce02e5ed81bf-kube-api-access-nhb2k\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189316 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189329 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189338 4904 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189348 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189356 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189366 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189376 4904 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e831e54b-3826-4415-9b43-ce02e5ed81bf-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189385 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189394 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189406 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189418 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189429 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.189439 4904 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e831e54b-3826-4415-9b43-ce02e5ed81bf-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.190929 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e8cddb62-7172-45e1-b162-ae9360043f92-audit-dir\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.192434 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-cliconfig\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.193280 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-session\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.193740 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e8cddb62-7172-45e1-b162-ae9360043f92-audit-policies\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.191994 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.194854 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-user-template-login\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.195202 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.195638 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-service-ca\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.196383 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-serving-cert\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.196473 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-system-router-certs\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.197117 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-user-template-error\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.200363 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.207608 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e8cddb62-7172-45e1-b162-ae9360043f92-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.209628 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5nm5\" (UniqueName: \"kubernetes.io/projected/e8cddb62-7172-45e1-b162-ae9360043f92-kube-api-access-x5nm5\") pod \"oauth-openshift-69b55d54f6-bnplh\" (UID: \"e8cddb62-7172-45e1-b162-ae9360043f92\") " pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.340671 4904 generic.go:334] "Generic (PLEG): container finished" podID="e3c0e6e8-ac12-4225-b17f-095591f8d2a7" containerID="5da3cbeae9ae533a4d90b170ce4554ef091abf4ef9159d7177c1ee4c266c4cbe" exitCode=0 Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.340802 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xphhn" event={"ID":"e3c0e6e8-ac12-4225-b17f-095591f8d2a7","Type":"ContainerDied","Data":"5da3cbeae9ae533a4d90b170ce4554ef091abf4ef9159d7177c1ee4c266c4cbe"} Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.343154 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.343478 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bfn42" event={"ID":"e831e54b-3826-4415-9b43-ce02e5ed81bf","Type":"ContainerDied","Data":"84274d7ba89098ae8f02f769ec2f990db221ec0183ffe85b4292f4c5e2e92e3e"} Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.343516 4904 scope.go:117] "RemoveContainer" containerID="0a42a22bc2927f70424cd130fee97a557db035a17cdc29afff69644a2db61987" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.370921 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bfn42"] Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.376189 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bfn42"] Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.383731 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.389009 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-977kk" podStartSLOduration=5.980908794 podStartE2EDuration="1m1.388992979s" podCreationTimestamp="2026-02-14 11:12:37 +0000 UTC" firstStartedPulling="2026-02-14 11:12:41.615646389 +0000 UTC m=+152.428411050" lastFinishedPulling="2026-02-14 11:13:37.023730564 +0000 UTC m=+207.836495235" observedRunningTime="2026-02-14 11:13:38.388141755 +0000 UTC m=+209.200906416" watchObservedRunningTime="2026-02-14 11:13:38.388992979 +0000 UTC m=+209.201757640" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.415794 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-977kk" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.416036 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-977kk" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.569745 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-69b55d54f6-bnplh"] Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.738095 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wlsqb" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.738385 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wlsqb" Feb 14 11:13:38 crc kubenswrapper[4904]: I0214 11:13:38.785041 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wlsqb" Feb 14 11:13:39 crc kubenswrapper[4904]: I0214 11:13:39.349879 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" event={"ID":"e8cddb62-7172-45e1-b162-ae9360043f92","Type":"ContainerStarted","Data":"379ca142ca89b2c2d3241727490ae47c1de5289beee52e42e707b1b2adf50505"} Feb 14 11:13:39 crc kubenswrapper[4904]: I0214 11:13:39.349920 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" event={"ID":"e8cddb62-7172-45e1-b162-ae9360043f92","Type":"ContainerStarted","Data":"a6c74b4a73ec1ca7a40a8733bf0bf34df48c137de2e2fc21563fe69b9dd6978c"} Feb 14 11:13:39 crc kubenswrapper[4904]: I0214 11:13:39.351050 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:39 crc kubenswrapper[4904]: I0214 11:13:39.354544 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xphhn" event={"ID":"e3c0e6e8-ac12-4225-b17f-095591f8d2a7","Type":"ContainerStarted","Data":"81cd3347cffd254b431b27e87f66632aebaf6177a6e642920516c57b2a12f0e7"} Feb 14 11:13:39 crc kubenswrapper[4904]: I0214 11:13:39.371218 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" podStartSLOduration=27.371203199 podStartE2EDuration="27.371203199s" podCreationTimestamp="2026-02-14 11:13:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:13:39.369221364 +0000 UTC m=+210.181986025" watchObservedRunningTime="2026-02-14 11:13:39.371203199 +0000 UTC m=+210.183967860" Feb 14 11:13:39 crc kubenswrapper[4904]: I0214 11:13:39.388821 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xphhn" podStartSLOduration=3.467466455 podStartE2EDuration="58.388801714s" podCreationTimestamp="2026-02-14 11:12:41 +0000 UTC" firstStartedPulling="2026-02-14 11:12:43.814900218 +0000 UTC m=+154.627664879" lastFinishedPulling="2026-02-14 11:13:38.736235477 +0000 UTC m=+209.549000138" observedRunningTime="2026-02-14 11:13:39.386782827 +0000 UTC m=+210.199547508" watchObservedRunningTime="2026-02-14 11:13:39.388801714 +0000 UTC m=+210.201566375" Feb 14 11:13:39 crc kubenswrapper[4904]: I0214 11:13:39.404413 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wlsqb" Feb 14 11:13:39 crc kubenswrapper[4904]: I0214 11:13:39.455392 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-977kk" podUID="92dc9e32-c069-4de6-a3ae-28b80b60d722" containerName="registry-server" probeResult="failure" output=< Feb 14 11:13:39 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 11:13:39 crc kubenswrapper[4904]: > Feb 14 11:13:39 crc kubenswrapper[4904]: I0214 11:13:39.749013 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-69b55d54f6-bnplh" Feb 14 11:13:39 crc kubenswrapper[4904]: I0214 11:13:39.843025 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e831e54b-3826-4415-9b43-ce02e5ed81bf" path="/var/lib/kubelet/pods/e831e54b-3826-4415-9b43-ce02e5ed81bf/volumes" Feb 14 11:13:40 crc kubenswrapper[4904]: I0214 11:13:40.673424 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-br8st" Feb 14 11:13:40 crc kubenswrapper[4904]: I0214 11:13:40.673499 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-br8st" Feb 14 11:13:40 crc kubenswrapper[4904]: I0214 11:13:40.719140 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-br8st" Feb 14 11:13:41 crc kubenswrapper[4904]: I0214 11:13:41.415170 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-br8st" Feb 14 11:13:41 crc kubenswrapper[4904]: I0214 11:13:41.723985 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xphhn" Feb 14 11:13:41 crc kubenswrapper[4904]: I0214 11:13:41.724060 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xphhn" Feb 14 11:13:42 crc kubenswrapper[4904]: I0214 11:13:42.650564 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wlsqb"] Feb 14 11:13:42 crc kubenswrapper[4904]: I0214 11:13:42.651120 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wlsqb" podUID="28f09505-e5c3-4105-88d5-9a6d5ed033e9" containerName="registry-server" containerID="cri-o://778442d1f9bdc3c507a90be7b2083d3b93f728b5a9e12f99fef04a8a5601fc6a" gracePeriod=2 Feb 14 11:13:42 crc kubenswrapper[4904]: I0214 11:13:42.782754 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xphhn" podUID="e3c0e6e8-ac12-4225-b17f-095591f8d2a7" containerName="registry-server" probeResult="failure" output=< Feb 14 11:13:42 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 11:13:42 crc kubenswrapper[4904]: > Feb 14 11:13:43 crc kubenswrapper[4904]: I0214 11:13:43.375170 4904 generic.go:334] "Generic (PLEG): container finished" podID="28f09505-e5c3-4105-88d5-9a6d5ed033e9" containerID="778442d1f9bdc3c507a90be7b2083d3b93f728b5a9e12f99fef04a8a5601fc6a" exitCode=0 Feb 14 11:13:43 crc kubenswrapper[4904]: I0214 11:13:43.375349 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlsqb" event={"ID":"28f09505-e5c3-4105-88d5-9a6d5ed033e9","Type":"ContainerDied","Data":"778442d1f9bdc3c507a90be7b2083d3b93f728b5a9e12f99fef04a8a5601fc6a"} Feb 14 11:13:43 crc kubenswrapper[4904]: I0214 11:13:43.375534 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlsqb" event={"ID":"28f09505-e5c3-4105-88d5-9a6d5ed033e9","Type":"ContainerDied","Data":"c68ee9c351a9fa1a2d72ec1053e093506230ba6c2190e3e478573e851c2ae420"} Feb 14 11:13:43 crc kubenswrapper[4904]: I0214 11:13:43.375554 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c68ee9c351a9fa1a2d72ec1053e093506230ba6c2190e3e478573e851c2ae420" Feb 14 11:13:43 crc kubenswrapper[4904]: I0214 11:13:43.378094 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlsqb" Feb 14 11:13:43 crc kubenswrapper[4904]: I0214 11:13:43.480963 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28f09505-e5c3-4105-88d5-9a6d5ed033e9-utilities\") pod \"28f09505-e5c3-4105-88d5-9a6d5ed033e9\" (UID: \"28f09505-e5c3-4105-88d5-9a6d5ed033e9\") " Feb 14 11:13:43 crc kubenswrapper[4904]: I0214 11:13:43.481035 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28f09505-e5c3-4105-88d5-9a6d5ed033e9-catalog-content\") pod \"28f09505-e5c3-4105-88d5-9a6d5ed033e9\" (UID: \"28f09505-e5c3-4105-88d5-9a6d5ed033e9\") " Feb 14 11:13:43 crc kubenswrapper[4904]: I0214 11:13:43.481070 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5x2x\" (UniqueName: \"kubernetes.io/projected/28f09505-e5c3-4105-88d5-9a6d5ed033e9-kube-api-access-h5x2x\") pod \"28f09505-e5c3-4105-88d5-9a6d5ed033e9\" (UID: \"28f09505-e5c3-4105-88d5-9a6d5ed033e9\") " Feb 14 11:13:43 crc kubenswrapper[4904]: I0214 11:13:43.481689 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28f09505-e5c3-4105-88d5-9a6d5ed033e9-utilities" (OuterVolumeSpecName: "utilities") pod "28f09505-e5c3-4105-88d5-9a6d5ed033e9" (UID: "28f09505-e5c3-4105-88d5-9a6d5ed033e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:13:43 crc kubenswrapper[4904]: I0214 11:13:43.487234 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28f09505-e5c3-4105-88d5-9a6d5ed033e9-kube-api-access-h5x2x" (OuterVolumeSpecName: "kube-api-access-h5x2x") pod "28f09505-e5c3-4105-88d5-9a6d5ed033e9" (UID: "28f09505-e5c3-4105-88d5-9a6d5ed033e9"). InnerVolumeSpecName "kube-api-access-h5x2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:13:43 crc kubenswrapper[4904]: I0214 11:13:43.527183 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28f09505-e5c3-4105-88d5-9a6d5ed033e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "28f09505-e5c3-4105-88d5-9a6d5ed033e9" (UID: "28f09505-e5c3-4105-88d5-9a6d5ed033e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:13:43 crc kubenswrapper[4904]: I0214 11:13:43.582692 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28f09505-e5c3-4105-88d5-9a6d5ed033e9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:43 crc kubenswrapper[4904]: I0214 11:13:43.582735 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5x2x\" (UniqueName: \"kubernetes.io/projected/28f09505-e5c3-4105-88d5-9a6d5ed033e9-kube-api-access-h5x2x\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:43 crc kubenswrapper[4904]: I0214 11:13:43.582750 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28f09505-e5c3-4105-88d5-9a6d5ed033e9-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:44 crc kubenswrapper[4904]: I0214 11:13:44.380277 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlsqb" Feb 14 11:13:44 crc kubenswrapper[4904]: I0214 11:13:44.398532 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wlsqb"] Feb 14 11:13:44 crc kubenswrapper[4904]: I0214 11:13:44.404937 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wlsqb"] Feb 14 11:13:44 crc kubenswrapper[4904]: I0214 11:13:44.856000 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-br8st"] Feb 14 11:13:44 crc kubenswrapper[4904]: I0214 11:13:44.856232 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-br8st" podUID="f5692d85-aa82-486b-889e-a8964e50b538" containerName="registry-server" containerID="cri-o://9f129083f5d4e701e483dedf6c257b7b435696006cdf5aa9c0dadfe1fa3d9d04" gracePeriod=2 Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.187812 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-br8st" Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.304634 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kf99\" (UniqueName: \"kubernetes.io/projected/f5692d85-aa82-486b-889e-a8964e50b538-kube-api-access-2kf99\") pod \"f5692d85-aa82-486b-889e-a8964e50b538\" (UID: \"f5692d85-aa82-486b-889e-a8964e50b538\") " Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.304976 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5692d85-aa82-486b-889e-a8964e50b538-catalog-content\") pod \"f5692d85-aa82-486b-889e-a8964e50b538\" (UID: \"f5692d85-aa82-486b-889e-a8964e50b538\") " Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.305057 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5692d85-aa82-486b-889e-a8964e50b538-utilities\") pod \"f5692d85-aa82-486b-889e-a8964e50b538\" (UID: \"f5692d85-aa82-486b-889e-a8964e50b538\") " Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.305945 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5692d85-aa82-486b-889e-a8964e50b538-utilities" (OuterVolumeSpecName: "utilities") pod "f5692d85-aa82-486b-889e-a8964e50b538" (UID: "f5692d85-aa82-486b-889e-a8964e50b538"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.321082 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5692d85-aa82-486b-889e-a8964e50b538-kube-api-access-2kf99" (OuterVolumeSpecName: "kube-api-access-2kf99") pod "f5692d85-aa82-486b-889e-a8964e50b538" (UID: "f5692d85-aa82-486b-889e-a8964e50b538"). InnerVolumeSpecName "kube-api-access-2kf99". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.339089 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5692d85-aa82-486b-889e-a8964e50b538-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5692d85-aa82-486b-889e-a8964e50b538" (UID: "f5692d85-aa82-486b-889e-a8964e50b538"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.390380 4904 generic.go:334] "Generic (PLEG): container finished" podID="f5692d85-aa82-486b-889e-a8964e50b538" containerID="9f129083f5d4e701e483dedf6c257b7b435696006cdf5aa9c0dadfe1fa3d9d04" exitCode=0 Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.390431 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-br8st" event={"ID":"f5692d85-aa82-486b-889e-a8964e50b538","Type":"ContainerDied","Data":"9f129083f5d4e701e483dedf6c257b7b435696006cdf5aa9c0dadfe1fa3d9d04"} Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.390470 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-br8st" event={"ID":"f5692d85-aa82-486b-889e-a8964e50b538","Type":"ContainerDied","Data":"6bb6191767325a72c03b131b48bdf7e32e898e35e52da6e26bc10e3abdf80da0"} Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.390497 4904 scope.go:117] "RemoveContainer" containerID="9f129083f5d4e701e483dedf6c257b7b435696006cdf5aa9c0dadfe1fa3d9d04" Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.390531 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-br8st" Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.406045 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5692d85-aa82-486b-889e-a8964e50b538-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.406067 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kf99\" (UniqueName: \"kubernetes.io/projected/f5692d85-aa82-486b-889e-a8964e50b538-kube-api-access-2kf99\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.406077 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5692d85-aa82-486b-889e-a8964e50b538-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.417744 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-br8st"] Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.418040 4904 scope.go:117] "RemoveContainer" containerID="b246c75b12d3a74095e39729b2a13c83b79500ec4e92b1f9ddda3a5ab73ba2df" Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.421164 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-br8st"] Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.430968 4904 scope.go:117] "RemoveContainer" containerID="9991f3746f32eb38dfbe0661228622068ded07209f3e12a996c2451cf248147c" Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.447965 4904 scope.go:117] "RemoveContainer" containerID="9f129083f5d4e701e483dedf6c257b7b435696006cdf5aa9c0dadfe1fa3d9d04" Feb 14 11:13:45 crc kubenswrapper[4904]: E0214 11:13:45.448322 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f129083f5d4e701e483dedf6c257b7b435696006cdf5aa9c0dadfe1fa3d9d04\": container with ID starting with 9f129083f5d4e701e483dedf6c257b7b435696006cdf5aa9c0dadfe1fa3d9d04 not found: ID does not exist" containerID="9f129083f5d4e701e483dedf6c257b7b435696006cdf5aa9c0dadfe1fa3d9d04" Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.448352 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f129083f5d4e701e483dedf6c257b7b435696006cdf5aa9c0dadfe1fa3d9d04"} err="failed to get container status \"9f129083f5d4e701e483dedf6c257b7b435696006cdf5aa9c0dadfe1fa3d9d04\": rpc error: code = NotFound desc = could not find container \"9f129083f5d4e701e483dedf6c257b7b435696006cdf5aa9c0dadfe1fa3d9d04\": container with ID starting with 9f129083f5d4e701e483dedf6c257b7b435696006cdf5aa9c0dadfe1fa3d9d04 not found: ID does not exist" Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.448375 4904 scope.go:117] "RemoveContainer" containerID="b246c75b12d3a74095e39729b2a13c83b79500ec4e92b1f9ddda3a5ab73ba2df" Feb 14 11:13:45 crc kubenswrapper[4904]: E0214 11:13:45.448732 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b246c75b12d3a74095e39729b2a13c83b79500ec4e92b1f9ddda3a5ab73ba2df\": container with ID starting with b246c75b12d3a74095e39729b2a13c83b79500ec4e92b1f9ddda3a5ab73ba2df not found: ID does not exist" containerID="b246c75b12d3a74095e39729b2a13c83b79500ec4e92b1f9ddda3a5ab73ba2df" Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.448782 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b246c75b12d3a74095e39729b2a13c83b79500ec4e92b1f9ddda3a5ab73ba2df"} err="failed to get container status \"b246c75b12d3a74095e39729b2a13c83b79500ec4e92b1f9ddda3a5ab73ba2df\": rpc error: code = NotFound desc = could not find container \"b246c75b12d3a74095e39729b2a13c83b79500ec4e92b1f9ddda3a5ab73ba2df\": container with ID starting with b246c75b12d3a74095e39729b2a13c83b79500ec4e92b1f9ddda3a5ab73ba2df not found: ID does not exist" Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.448816 4904 scope.go:117] "RemoveContainer" containerID="9991f3746f32eb38dfbe0661228622068ded07209f3e12a996c2451cf248147c" Feb 14 11:13:45 crc kubenswrapper[4904]: E0214 11:13:45.449312 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9991f3746f32eb38dfbe0661228622068ded07209f3e12a996c2451cf248147c\": container with ID starting with 9991f3746f32eb38dfbe0661228622068ded07209f3e12a996c2451cf248147c not found: ID does not exist" containerID="9991f3746f32eb38dfbe0661228622068ded07209f3e12a996c2451cf248147c" Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.449334 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9991f3746f32eb38dfbe0661228622068ded07209f3e12a996c2451cf248147c"} err="failed to get container status \"9991f3746f32eb38dfbe0661228622068ded07209f3e12a996c2451cf248147c\": rpc error: code = NotFound desc = could not find container \"9991f3746f32eb38dfbe0661228622068ded07209f3e12a996c2451cf248147c\": container with ID starting with 9991f3746f32eb38dfbe0661228622068ded07209f3e12a996c2451cf248147c not found: ID does not exist" Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.843321 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28f09505-e5c3-4105-88d5-9a6d5ed033e9" path="/var/lib/kubelet/pods/28f09505-e5c3-4105-88d5-9a6d5ed033e9/volumes" Feb 14 11:13:45 crc kubenswrapper[4904]: I0214 11:13:45.843962 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5692d85-aa82-486b-889e-a8964e50b538" path="/var/lib/kubelet/pods/f5692d85-aa82-486b-889e-a8964e50b538/volumes" Feb 14 11:13:46 crc kubenswrapper[4904]: I0214 11:13:46.382996 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:13:46 crc kubenswrapper[4904]: I0214 11:13:46.383071 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:13:46 crc kubenswrapper[4904]: I0214 11:13:46.383141 4904 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:13:46 crc kubenswrapper[4904]: I0214 11:13:46.383884 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228"} pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 14 11:13:46 crc kubenswrapper[4904]: I0214 11:13:46.383970 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" containerID="cri-o://29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228" gracePeriod=600 Feb 14 11:13:47 crc kubenswrapper[4904]: I0214 11:13:47.404746 4904 generic.go:334] "Generic (PLEG): container finished" podID="7790d276-2506-4641-8c22-67e6441417f0" containerID="29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228" exitCode=0 Feb 14 11:13:47 crc kubenswrapper[4904]: I0214 11:13:47.404824 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerDied","Data":"29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228"} Feb 14 11:13:47 crc kubenswrapper[4904]: I0214 11:13:47.405107 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"903ea27dacf6775c4c9f46b3a1c3654b638e3bdce73cd947b4f81a7bea2eca49"} Feb 14 11:13:48 crc kubenswrapper[4904]: I0214 11:13:48.451391 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-977kk" Feb 14 11:13:48 crc kubenswrapper[4904]: I0214 11:13:48.492942 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-977kk" Feb 14 11:13:51 crc kubenswrapper[4904]: I0214 11:13:51.759221 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xphhn" Feb 14 11:13:51 crc kubenswrapper[4904]: I0214 11:13:51.802427 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xphhn" Feb 14 11:13:52 crc kubenswrapper[4904]: I0214 11:13:52.054830 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xphhn"] Feb 14 11:13:53 crc kubenswrapper[4904]: I0214 11:13:53.450644 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xphhn" podUID="e3c0e6e8-ac12-4225-b17f-095591f8d2a7" containerName="registry-server" containerID="cri-o://81cd3347cffd254b431b27e87f66632aebaf6177a6e642920516c57b2a12f0e7" gracePeriod=2 Feb 14 11:13:53 crc kubenswrapper[4904]: I0214 11:13:53.838614 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xphhn" Feb 14 11:13:53 crc kubenswrapper[4904]: I0214 11:13:53.922718 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3c0e6e8-ac12-4225-b17f-095591f8d2a7-catalog-content\") pod \"e3c0e6e8-ac12-4225-b17f-095591f8d2a7\" (UID: \"e3c0e6e8-ac12-4225-b17f-095591f8d2a7\") " Feb 14 11:13:53 crc kubenswrapper[4904]: I0214 11:13:53.922825 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jcv6\" (UniqueName: \"kubernetes.io/projected/e3c0e6e8-ac12-4225-b17f-095591f8d2a7-kube-api-access-8jcv6\") pod \"e3c0e6e8-ac12-4225-b17f-095591f8d2a7\" (UID: \"e3c0e6e8-ac12-4225-b17f-095591f8d2a7\") " Feb 14 11:13:53 crc kubenswrapper[4904]: I0214 11:13:53.922894 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3c0e6e8-ac12-4225-b17f-095591f8d2a7-utilities\") pod \"e3c0e6e8-ac12-4225-b17f-095591f8d2a7\" (UID: \"e3c0e6e8-ac12-4225-b17f-095591f8d2a7\") " Feb 14 11:13:53 crc kubenswrapper[4904]: I0214 11:13:53.924014 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3c0e6e8-ac12-4225-b17f-095591f8d2a7-utilities" (OuterVolumeSpecName: "utilities") pod "e3c0e6e8-ac12-4225-b17f-095591f8d2a7" (UID: "e3c0e6e8-ac12-4225-b17f-095591f8d2a7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:13:53 crc kubenswrapper[4904]: I0214 11:13:53.924205 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3c0e6e8-ac12-4225-b17f-095591f8d2a7-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:53 crc kubenswrapper[4904]: I0214 11:13:53.948806 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3c0e6e8-ac12-4225-b17f-095591f8d2a7-kube-api-access-8jcv6" (OuterVolumeSpecName: "kube-api-access-8jcv6") pod "e3c0e6e8-ac12-4225-b17f-095591f8d2a7" (UID: "e3c0e6e8-ac12-4225-b17f-095591f8d2a7"). InnerVolumeSpecName "kube-api-access-8jcv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:13:54 crc kubenswrapper[4904]: I0214 11:13:54.025922 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jcv6\" (UniqueName: \"kubernetes.io/projected/e3c0e6e8-ac12-4225-b17f-095591f8d2a7-kube-api-access-8jcv6\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:54 crc kubenswrapper[4904]: I0214 11:13:54.052322 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3c0e6e8-ac12-4225-b17f-095591f8d2a7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e3c0e6e8-ac12-4225-b17f-095591f8d2a7" (UID: "e3c0e6e8-ac12-4225-b17f-095591f8d2a7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:13:54 crc kubenswrapper[4904]: I0214 11:13:54.126782 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3c0e6e8-ac12-4225-b17f-095591f8d2a7-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:13:54 crc kubenswrapper[4904]: I0214 11:13:54.458875 4904 generic.go:334] "Generic (PLEG): container finished" podID="e3c0e6e8-ac12-4225-b17f-095591f8d2a7" containerID="81cd3347cffd254b431b27e87f66632aebaf6177a6e642920516c57b2a12f0e7" exitCode=0 Feb 14 11:13:54 crc kubenswrapper[4904]: I0214 11:13:54.458926 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xphhn" event={"ID":"e3c0e6e8-ac12-4225-b17f-095591f8d2a7","Type":"ContainerDied","Data":"81cd3347cffd254b431b27e87f66632aebaf6177a6e642920516c57b2a12f0e7"} Feb 14 11:13:54 crc kubenswrapper[4904]: I0214 11:13:54.458960 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xphhn" event={"ID":"e3c0e6e8-ac12-4225-b17f-095591f8d2a7","Type":"ContainerDied","Data":"00750ee463fe5d80a8b5aaf2129de84ec9145448f7c7200c7ddf9797f8b8f6dc"} Feb 14 11:13:54 crc kubenswrapper[4904]: I0214 11:13:54.458977 4904 scope.go:117] "RemoveContainer" containerID="81cd3347cffd254b431b27e87f66632aebaf6177a6e642920516c57b2a12f0e7" Feb 14 11:13:54 crc kubenswrapper[4904]: I0214 11:13:54.459045 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xphhn" Feb 14 11:13:54 crc kubenswrapper[4904]: I0214 11:13:54.475651 4904 scope.go:117] "RemoveContainer" containerID="5da3cbeae9ae533a4d90b170ce4554ef091abf4ef9159d7177c1ee4c266c4cbe" Feb 14 11:13:54 crc kubenswrapper[4904]: I0214 11:13:54.484925 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xphhn"] Feb 14 11:13:54 crc kubenswrapper[4904]: I0214 11:13:54.487486 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xphhn"] Feb 14 11:13:54 crc kubenswrapper[4904]: I0214 11:13:54.506134 4904 scope.go:117] "RemoveContainer" containerID="c9b142bf1889a0c950f580730dbf830044d28ca073d00afc7f104d0ec4342fc0" Feb 14 11:13:54 crc kubenswrapper[4904]: I0214 11:13:54.519526 4904 scope.go:117] "RemoveContainer" containerID="81cd3347cffd254b431b27e87f66632aebaf6177a6e642920516c57b2a12f0e7" Feb 14 11:13:54 crc kubenswrapper[4904]: E0214 11:13:54.519950 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81cd3347cffd254b431b27e87f66632aebaf6177a6e642920516c57b2a12f0e7\": container with ID starting with 81cd3347cffd254b431b27e87f66632aebaf6177a6e642920516c57b2a12f0e7 not found: ID does not exist" containerID="81cd3347cffd254b431b27e87f66632aebaf6177a6e642920516c57b2a12f0e7" Feb 14 11:13:54 crc kubenswrapper[4904]: I0214 11:13:54.519983 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81cd3347cffd254b431b27e87f66632aebaf6177a6e642920516c57b2a12f0e7"} err="failed to get container status \"81cd3347cffd254b431b27e87f66632aebaf6177a6e642920516c57b2a12f0e7\": rpc error: code = NotFound desc = could not find container \"81cd3347cffd254b431b27e87f66632aebaf6177a6e642920516c57b2a12f0e7\": container with ID starting with 81cd3347cffd254b431b27e87f66632aebaf6177a6e642920516c57b2a12f0e7 not found: ID does not exist" Feb 14 11:13:54 crc kubenswrapper[4904]: I0214 11:13:54.520009 4904 scope.go:117] "RemoveContainer" containerID="5da3cbeae9ae533a4d90b170ce4554ef091abf4ef9159d7177c1ee4c266c4cbe" Feb 14 11:13:54 crc kubenswrapper[4904]: E0214 11:13:54.520258 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5da3cbeae9ae533a4d90b170ce4554ef091abf4ef9159d7177c1ee4c266c4cbe\": container with ID starting with 5da3cbeae9ae533a4d90b170ce4554ef091abf4ef9159d7177c1ee4c266c4cbe not found: ID does not exist" containerID="5da3cbeae9ae533a4d90b170ce4554ef091abf4ef9159d7177c1ee4c266c4cbe" Feb 14 11:13:54 crc kubenswrapper[4904]: I0214 11:13:54.520307 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5da3cbeae9ae533a4d90b170ce4554ef091abf4ef9159d7177c1ee4c266c4cbe"} err="failed to get container status \"5da3cbeae9ae533a4d90b170ce4554ef091abf4ef9159d7177c1ee4c266c4cbe\": rpc error: code = NotFound desc = could not find container \"5da3cbeae9ae533a4d90b170ce4554ef091abf4ef9159d7177c1ee4c266c4cbe\": container with ID starting with 5da3cbeae9ae533a4d90b170ce4554ef091abf4ef9159d7177c1ee4c266c4cbe not found: ID does not exist" Feb 14 11:13:54 crc kubenswrapper[4904]: I0214 11:13:54.520322 4904 scope.go:117] "RemoveContainer" containerID="c9b142bf1889a0c950f580730dbf830044d28ca073d00afc7f104d0ec4342fc0" Feb 14 11:13:54 crc kubenswrapper[4904]: E0214 11:13:54.520534 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9b142bf1889a0c950f580730dbf830044d28ca073d00afc7f104d0ec4342fc0\": container with ID starting with c9b142bf1889a0c950f580730dbf830044d28ca073d00afc7f104d0ec4342fc0 not found: ID does not exist" containerID="c9b142bf1889a0c950f580730dbf830044d28ca073d00afc7f104d0ec4342fc0" Feb 14 11:13:54 crc kubenswrapper[4904]: I0214 11:13:54.520556 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9b142bf1889a0c950f580730dbf830044d28ca073d00afc7f104d0ec4342fc0"} err="failed to get container status \"c9b142bf1889a0c950f580730dbf830044d28ca073d00afc7f104d0ec4342fc0\": rpc error: code = NotFound desc = could not find container \"c9b142bf1889a0c950f580730dbf830044d28ca073d00afc7f104d0ec4342fc0\": container with ID starting with c9b142bf1889a0c950f580730dbf830044d28ca073d00afc7f104d0ec4342fc0 not found: ID does not exist" Feb 14 11:13:55 crc kubenswrapper[4904]: I0214 11:13:55.846337 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3c0e6e8-ac12-4225-b17f-095591f8d2a7" path="/var/lib/kubelet/pods/e3c0e6e8-ac12-4225-b17f-095591f8d2a7/volumes" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.704331 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-977kk"] Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.705303 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-977kk" podUID="92dc9e32-c069-4de6-a3ae-28b80b60d722" containerName="registry-server" containerID="cri-o://62d391f394fced8cef03cbba49f457f6818005d1c190e5ca767d45bbc9d57a2c" gracePeriod=30 Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.720082 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bvpm5"] Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.720558 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bvpm5" podUID="63c9ce1a-01e3-42f3-8027-bfda9be0b3ff" containerName="registry-server" containerID="cri-o://6c243d2633d6cd00022d141ab70d97df2920850cf29e1c198bbc7684c36ac17f" gracePeriod=30 Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.723656 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9pzkg"] Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.723868 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" podUID="f130a8a5-24b3-4916-a0c7-b1fa5e55cad4" containerName="marketplace-operator" containerID="cri-o://c58cdbc463ae1de3ed0c35e320580faeaddb6a95577c99c69fe14ae19e9381c9" gracePeriod=30 Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.726388 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nvpkp"] Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.726698 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nvpkp" podUID="94b8c7dc-4ce9-4434-a387-e19139fe8a3d" containerName="registry-server" containerID="cri-o://91278d384c32444678079c8fda4688f16a28bc0032d085b4a9216e1b1801b357" gracePeriod=30 Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.751537 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-h5v26"] Feb 14 11:14:00 crc kubenswrapper[4904]: E0214 11:14:00.751747 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3c0e6e8-ac12-4225-b17f-095591f8d2a7" containerName="registry-server" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.751763 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3c0e6e8-ac12-4225-b17f-095591f8d2a7" containerName="registry-server" Feb 14 11:14:00 crc kubenswrapper[4904]: E0214 11:14:00.751780 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5692d85-aa82-486b-889e-a8964e50b538" containerName="extract-content" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.751787 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5692d85-aa82-486b-889e-a8964e50b538" containerName="extract-content" Feb 14 11:14:00 crc kubenswrapper[4904]: E0214 11:14:00.751798 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28f09505-e5c3-4105-88d5-9a6d5ed033e9" containerName="registry-server" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.751805 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="28f09505-e5c3-4105-88d5-9a6d5ed033e9" containerName="registry-server" Feb 14 11:14:00 crc kubenswrapper[4904]: E0214 11:14:00.751815 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5692d85-aa82-486b-889e-a8964e50b538" containerName="registry-server" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.751822 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5692d85-aa82-486b-889e-a8964e50b538" containerName="registry-server" Feb 14 11:14:00 crc kubenswrapper[4904]: E0214 11:14:00.751834 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3c0e6e8-ac12-4225-b17f-095591f8d2a7" containerName="extract-content" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.751840 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3c0e6e8-ac12-4225-b17f-095591f8d2a7" containerName="extract-content" Feb 14 11:14:00 crc kubenswrapper[4904]: E0214 11:14:00.751915 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28f09505-e5c3-4105-88d5-9a6d5ed033e9" containerName="extract-utilities" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.751921 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="28f09505-e5c3-4105-88d5-9a6d5ed033e9" containerName="extract-utilities" Feb 14 11:14:00 crc kubenswrapper[4904]: E0214 11:14:00.751930 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5692d85-aa82-486b-889e-a8964e50b538" containerName="extract-utilities" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.751935 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5692d85-aa82-486b-889e-a8964e50b538" containerName="extract-utilities" Feb 14 11:14:00 crc kubenswrapper[4904]: E0214 11:14:00.751942 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28f09505-e5c3-4105-88d5-9a6d5ed033e9" containerName="extract-content" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.751949 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="28f09505-e5c3-4105-88d5-9a6d5ed033e9" containerName="extract-content" Feb 14 11:14:00 crc kubenswrapper[4904]: E0214 11:14:00.751960 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3c0e6e8-ac12-4225-b17f-095591f8d2a7" containerName="extract-utilities" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.751965 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3c0e6e8-ac12-4225-b17f-095591f8d2a7" containerName="extract-utilities" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.752076 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5692d85-aa82-486b-889e-a8964e50b538" containerName="registry-server" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.752093 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="28f09505-e5c3-4105-88d5-9a6d5ed033e9" containerName="registry-server" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.752103 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3c0e6e8-ac12-4225-b17f-095591f8d2a7" containerName="registry-server" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.752422 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-krh2c"] Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.752532 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-h5v26" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.753021 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-krh2c" podUID="49e91378-b5ff-4480-abbf-765ccc061978" containerName="registry-server" containerID="cri-o://57d5b7a6e0f85a669f18f54f74eccce27d403d2021bf46561a6cacc13cc5beb6" gracePeriod=30 Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.815036 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2104a707-8d9b-43e6-9c74-f41bdafcf818-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-h5v26\" (UID: \"2104a707-8d9b-43e6-9c74-f41bdafcf818\") " pod="openshift-marketplace/marketplace-operator-79b997595-h5v26" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.815109 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2104a707-8d9b-43e6-9c74-f41bdafcf818-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-h5v26\" (UID: \"2104a707-8d9b-43e6-9c74-f41bdafcf818\") " pod="openshift-marketplace/marketplace-operator-79b997595-h5v26" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.815153 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6sdg\" (UniqueName: \"kubernetes.io/projected/2104a707-8d9b-43e6-9c74-f41bdafcf818-kube-api-access-t6sdg\") pod \"marketplace-operator-79b997595-h5v26\" (UID: \"2104a707-8d9b-43e6-9c74-f41bdafcf818\") " pod="openshift-marketplace/marketplace-operator-79b997595-h5v26" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.836627 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-h5v26"] Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.877708 4904 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9pzkg container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": read tcp 10.217.0.2:48396->10.217.0.34:8080: read: connection reset by peer" start-of-body= Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.877764 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" podUID="f130a8a5-24b3-4916-a0c7-b1fa5e55cad4" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": read tcp 10.217.0.2:48396->10.217.0.34:8080: read: connection reset by peer" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.919560 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6sdg\" (UniqueName: \"kubernetes.io/projected/2104a707-8d9b-43e6-9c74-f41bdafcf818-kube-api-access-t6sdg\") pod \"marketplace-operator-79b997595-h5v26\" (UID: \"2104a707-8d9b-43e6-9c74-f41bdafcf818\") " pod="openshift-marketplace/marketplace-operator-79b997595-h5v26" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.919651 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2104a707-8d9b-43e6-9c74-f41bdafcf818-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-h5v26\" (UID: \"2104a707-8d9b-43e6-9c74-f41bdafcf818\") " pod="openshift-marketplace/marketplace-operator-79b997595-h5v26" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.919707 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2104a707-8d9b-43e6-9c74-f41bdafcf818-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-h5v26\" (UID: \"2104a707-8d9b-43e6-9c74-f41bdafcf818\") " pod="openshift-marketplace/marketplace-operator-79b997595-h5v26" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.921676 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2104a707-8d9b-43e6-9c74-f41bdafcf818-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-h5v26\" (UID: \"2104a707-8d9b-43e6-9c74-f41bdafcf818\") " pod="openshift-marketplace/marketplace-operator-79b997595-h5v26" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.932310 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2104a707-8d9b-43e6-9c74-f41bdafcf818-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-h5v26\" (UID: \"2104a707-8d9b-43e6-9c74-f41bdafcf818\") " pod="openshift-marketplace/marketplace-operator-79b997595-h5v26" Feb 14 11:14:00 crc kubenswrapper[4904]: I0214 11:14:00.941608 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6sdg\" (UniqueName: \"kubernetes.io/projected/2104a707-8d9b-43e6-9c74-f41bdafcf818-kube-api-access-t6sdg\") pod \"marketplace-operator-79b997595-h5v26\" (UID: \"2104a707-8d9b-43e6-9c74-f41bdafcf818\") " pod="openshift-marketplace/marketplace-operator-79b997595-h5v26" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.068520 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-h5v26" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.135793 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-977kk" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.229425 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92dc9e32-c069-4de6-a3ae-28b80b60d722-catalog-content\") pod \"92dc9e32-c069-4de6-a3ae-28b80b60d722\" (UID: \"92dc9e32-c069-4de6-a3ae-28b80b60d722\") " Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.229529 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vw586\" (UniqueName: \"kubernetes.io/projected/92dc9e32-c069-4de6-a3ae-28b80b60d722-kube-api-access-vw586\") pod \"92dc9e32-c069-4de6-a3ae-28b80b60d722\" (UID: \"92dc9e32-c069-4de6-a3ae-28b80b60d722\") " Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.229599 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92dc9e32-c069-4de6-a3ae-28b80b60d722-utilities\") pod \"92dc9e32-c069-4de6-a3ae-28b80b60d722\" (UID: \"92dc9e32-c069-4de6-a3ae-28b80b60d722\") " Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.237950 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92dc9e32-c069-4de6-a3ae-28b80b60d722-kube-api-access-vw586" (OuterVolumeSpecName: "kube-api-access-vw586") pod "92dc9e32-c069-4de6-a3ae-28b80b60d722" (UID: "92dc9e32-c069-4de6-a3ae-28b80b60d722"). InnerVolumeSpecName "kube-api-access-vw586". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.238396 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92dc9e32-c069-4de6-a3ae-28b80b60d722-utilities" (OuterVolumeSpecName: "utilities") pod "92dc9e32-c069-4de6-a3ae-28b80b60d722" (UID: "92dc9e32-c069-4de6-a3ae-28b80b60d722"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.308188 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nvpkp" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.331726 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vw586\" (UniqueName: \"kubernetes.io/projected/92dc9e32-c069-4de6-a3ae-28b80b60d722-kube-api-access-vw586\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.331749 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92dc9e32-c069-4de6-a3ae-28b80b60d722-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.343044 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92dc9e32-c069-4de6-a3ae-28b80b60d722-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92dc9e32-c069-4de6-a3ae-28b80b60d722" (UID: "92dc9e32-c069-4de6-a3ae-28b80b60d722"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.370064 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bvpm5" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.379272 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krh2c" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.433898 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kh6f\" (UniqueName: \"kubernetes.io/projected/63c9ce1a-01e3-42f3-8027-bfda9be0b3ff-kube-api-access-6kh6f\") pod \"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff\" (UID: \"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff\") " Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.434001 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pv727\" (UniqueName: \"kubernetes.io/projected/49e91378-b5ff-4480-abbf-765ccc061978-kube-api-access-pv727\") pod \"49e91378-b5ff-4480-abbf-765ccc061978\" (UID: \"49e91378-b5ff-4480-abbf-765ccc061978\") " Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.434040 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94b8c7dc-4ce9-4434-a387-e19139fe8a3d-catalog-content\") pod \"94b8c7dc-4ce9-4434-a387-e19139fe8a3d\" (UID: \"94b8c7dc-4ce9-4434-a387-e19139fe8a3d\") " Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.434059 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49e91378-b5ff-4480-abbf-765ccc061978-utilities\") pod \"49e91378-b5ff-4480-abbf-765ccc061978\" (UID: \"49e91378-b5ff-4480-abbf-765ccc061978\") " Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.434074 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdhkw\" (UniqueName: \"kubernetes.io/projected/94b8c7dc-4ce9-4434-a387-e19139fe8a3d-kube-api-access-rdhkw\") pod \"94b8c7dc-4ce9-4434-a387-e19139fe8a3d\" (UID: \"94b8c7dc-4ce9-4434-a387-e19139fe8a3d\") " Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.434088 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49e91378-b5ff-4480-abbf-765ccc061978-catalog-content\") pod \"49e91378-b5ff-4480-abbf-765ccc061978\" (UID: \"49e91378-b5ff-4480-abbf-765ccc061978\") " Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.434164 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63c9ce1a-01e3-42f3-8027-bfda9be0b3ff-catalog-content\") pod \"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff\" (UID: \"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff\") " Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.434185 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63c9ce1a-01e3-42f3-8027-bfda9be0b3ff-utilities\") pod \"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff\" (UID: \"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff\") " Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.434198 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94b8c7dc-4ce9-4434-a387-e19139fe8a3d-utilities\") pod \"94b8c7dc-4ce9-4434-a387-e19139fe8a3d\" (UID: \"94b8c7dc-4ce9-4434-a387-e19139fe8a3d\") " Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.434404 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92dc9e32-c069-4de6-a3ae-28b80b60d722-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.435096 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94b8c7dc-4ce9-4434-a387-e19139fe8a3d-utilities" (OuterVolumeSpecName: "utilities") pod "94b8c7dc-4ce9-4434-a387-e19139fe8a3d" (UID: "94b8c7dc-4ce9-4434-a387-e19139fe8a3d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.437399 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63c9ce1a-01e3-42f3-8027-bfda9be0b3ff-utilities" (OuterVolumeSpecName: "utilities") pod "63c9ce1a-01e3-42f3-8027-bfda9be0b3ff" (UID: "63c9ce1a-01e3-42f3-8027-bfda9be0b3ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.439512 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49e91378-b5ff-4480-abbf-765ccc061978-utilities" (OuterVolumeSpecName: "utilities") pod "49e91378-b5ff-4480-abbf-765ccc061978" (UID: "49e91378-b5ff-4480-abbf-765ccc061978"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.443571 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94b8c7dc-4ce9-4434-a387-e19139fe8a3d-kube-api-access-rdhkw" (OuterVolumeSpecName: "kube-api-access-rdhkw") pod "94b8c7dc-4ce9-4434-a387-e19139fe8a3d" (UID: "94b8c7dc-4ce9-4434-a387-e19139fe8a3d"). InnerVolumeSpecName "kube-api-access-rdhkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.462538 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63c9ce1a-01e3-42f3-8027-bfda9be0b3ff-kube-api-access-6kh6f" (OuterVolumeSpecName: "kube-api-access-6kh6f") pod "63c9ce1a-01e3-42f3-8027-bfda9be0b3ff" (UID: "63c9ce1a-01e3-42f3-8027-bfda9be0b3ff"). InnerVolumeSpecName "kube-api-access-6kh6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.475661 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94b8c7dc-4ce9-4434-a387-e19139fe8a3d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94b8c7dc-4ce9-4434-a387-e19139fe8a3d" (UID: "94b8c7dc-4ce9-4434-a387-e19139fe8a3d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.476622 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49e91378-b5ff-4480-abbf-765ccc061978-kube-api-access-pv727" (OuterVolumeSpecName: "kube-api-access-pv727") pod "49e91378-b5ff-4480-abbf-765ccc061978" (UID: "49e91378-b5ff-4480-abbf-765ccc061978"). InnerVolumeSpecName "kube-api-access-pv727". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.536932 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63c9ce1a-01e3-42f3-8027-bfda9be0b3ff-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.536980 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94b8c7dc-4ce9-4434-a387-e19139fe8a3d-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.536994 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kh6f\" (UniqueName: \"kubernetes.io/projected/63c9ce1a-01e3-42f3-8027-bfda9be0b3ff-kube-api-access-6kh6f\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.537009 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pv727\" (UniqueName: \"kubernetes.io/projected/49e91378-b5ff-4480-abbf-765ccc061978-kube-api-access-pv727\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.537029 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94b8c7dc-4ce9-4434-a387-e19139fe8a3d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.537052 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49e91378-b5ff-4480-abbf-765ccc061978-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.537072 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdhkw\" (UniqueName: \"kubernetes.io/projected/94b8c7dc-4ce9-4434-a387-e19139fe8a3d-kube-api-access-rdhkw\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.544474 4904 generic.go:334] "Generic (PLEG): container finished" podID="49e91378-b5ff-4480-abbf-765ccc061978" containerID="57d5b7a6e0f85a669f18f54f74eccce27d403d2021bf46561a6cacc13cc5beb6" exitCode=0 Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.544628 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krh2c" event={"ID":"49e91378-b5ff-4480-abbf-765ccc061978","Type":"ContainerDied","Data":"57d5b7a6e0f85a669f18f54f74eccce27d403d2021bf46561a6cacc13cc5beb6"} Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.544663 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krh2c" event={"ID":"49e91378-b5ff-4480-abbf-765ccc061978","Type":"ContainerDied","Data":"01a2e06d56c934699ea02506921cc080998cfdf2a52e96dd13e3d62107da12e8"} Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.544682 4904 scope.go:117] "RemoveContainer" containerID="57d5b7a6e0f85a669f18f54f74eccce27d403d2021bf46561a6cacc13cc5beb6" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.544955 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krh2c" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.565745 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-h5v26"] Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.568732 4904 scope.go:117] "RemoveContainer" containerID="426f9681e2ee030a5145f3d68a2cbc69bb9e984aaa9607ede33fb69eb9d01750" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.570940 4904 generic.go:334] "Generic (PLEG): container finished" podID="92dc9e32-c069-4de6-a3ae-28b80b60d722" containerID="62d391f394fced8cef03cbba49f457f6818005d1c190e5ca767d45bbc9d57a2c" exitCode=0 Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.571085 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-977kk" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.571128 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-977kk" event={"ID":"92dc9e32-c069-4de6-a3ae-28b80b60d722","Type":"ContainerDied","Data":"62d391f394fced8cef03cbba49f457f6818005d1c190e5ca767d45bbc9d57a2c"} Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.571167 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-977kk" event={"ID":"92dc9e32-c069-4de6-a3ae-28b80b60d722","Type":"ContainerDied","Data":"fecbde0a237da81f7630e72ce2d0a2dc763ad16d5ac2f9bbe555882218609cfe"} Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.583966 4904 generic.go:334] "Generic (PLEG): container finished" podID="63c9ce1a-01e3-42f3-8027-bfda9be0b3ff" containerID="6c243d2633d6cd00022d141ab70d97df2920850cf29e1c198bbc7684c36ac17f" exitCode=0 Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.584067 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bvpm5" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.584446 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bvpm5" event={"ID":"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff","Type":"ContainerDied","Data":"6c243d2633d6cd00022d141ab70d97df2920850cf29e1c198bbc7684c36ac17f"} Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.590152 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bvpm5" event={"ID":"63c9ce1a-01e3-42f3-8027-bfda9be0b3ff","Type":"ContainerDied","Data":"a9cfcde1152608ae1472a0761cda82a2bf852954afbcd565db74b202ecf6632b"} Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.611903 4904 scope.go:117] "RemoveContainer" containerID="33ff4599352ab484361317e1141542245cd3827b024c49d1b281f4433e833f9c" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.619269 4904 generic.go:334] "Generic (PLEG): container finished" podID="f130a8a5-24b3-4916-a0c7-b1fa5e55cad4" containerID="c58cdbc463ae1de3ed0c35e320580faeaddb6a95577c99c69fe14ae19e9381c9" exitCode=0 Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.619330 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" event={"ID":"f130a8a5-24b3-4916-a0c7-b1fa5e55cad4","Type":"ContainerDied","Data":"c58cdbc463ae1de3ed0c35e320580faeaddb6a95577c99c69fe14ae19e9381c9"} Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.626975 4904 generic.go:334] "Generic (PLEG): container finished" podID="94b8c7dc-4ce9-4434-a387-e19139fe8a3d" containerID="91278d384c32444678079c8fda4688f16a28bc0032d085b4a9216e1b1801b357" exitCode=0 Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.627011 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nvpkp" event={"ID":"94b8c7dc-4ce9-4434-a387-e19139fe8a3d","Type":"ContainerDied","Data":"91278d384c32444678079c8fda4688f16a28bc0032d085b4a9216e1b1801b357"} Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.627037 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nvpkp" event={"ID":"94b8c7dc-4ce9-4434-a387-e19139fe8a3d","Type":"ContainerDied","Data":"782d6d853e618365d7d2209b9b4d5d1d260fc2606a84afc39e8d55e1473acaa4"} Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.627202 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nvpkp" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.645081 4904 scope.go:117] "RemoveContainer" containerID="57d5b7a6e0f85a669f18f54f74eccce27d403d2021bf46561a6cacc13cc5beb6" Feb 14 11:14:01 crc kubenswrapper[4904]: E0214 11:14:01.646157 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57d5b7a6e0f85a669f18f54f74eccce27d403d2021bf46561a6cacc13cc5beb6\": container with ID starting with 57d5b7a6e0f85a669f18f54f74eccce27d403d2021bf46561a6cacc13cc5beb6 not found: ID does not exist" containerID="57d5b7a6e0f85a669f18f54f74eccce27d403d2021bf46561a6cacc13cc5beb6" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.646232 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57d5b7a6e0f85a669f18f54f74eccce27d403d2021bf46561a6cacc13cc5beb6"} err="failed to get container status \"57d5b7a6e0f85a669f18f54f74eccce27d403d2021bf46561a6cacc13cc5beb6\": rpc error: code = NotFound desc = could not find container \"57d5b7a6e0f85a669f18f54f74eccce27d403d2021bf46561a6cacc13cc5beb6\": container with ID starting with 57d5b7a6e0f85a669f18f54f74eccce27d403d2021bf46561a6cacc13cc5beb6 not found: ID does not exist" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.646268 4904 scope.go:117] "RemoveContainer" containerID="426f9681e2ee030a5145f3d68a2cbc69bb9e984aaa9607ede33fb69eb9d01750" Feb 14 11:14:01 crc kubenswrapper[4904]: E0214 11:14:01.646953 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"426f9681e2ee030a5145f3d68a2cbc69bb9e984aaa9607ede33fb69eb9d01750\": container with ID starting with 426f9681e2ee030a5145f3d68a2cbc69bb9e984aaa9607ede33fb69eb9d01750 not found: ID does not exist" containerID="426f9681e2ee030a5145f3d68a2cbc69bb9e984aaa9607ede33fb69eb9d01750" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.646980 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"426f9681e2ee030a5145f3d68a2cbc69bb9e984aaa9607ede33fb69eb9d01750"} err="failed to get container status \"426f9681e2ee030a5145f3d68a2cbc69bb9e984aaa9607ede33fb69eb9d01750\": rpc error: code = NotFound desc = could not find container \"426f9681e2ee030a5145f3d68a2cbc69bb9e984aaa9607ede33fb69eb9d01750\": container with ID starting with 426f9681e2ee030a5145f3d68a2cbc69bb9e984aaa9607ede33fb69eb9d01750 not found: ID does not exist" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.647020 4904 scope.go:117] "RemoveContainer" containerID="33ff4599352ab484361317e1141542245cd3827b024c49d1b281f4433e833f9c" Feb 14 11:14:01 crc kubenswrapper[4904]: E0214 11:14:01.647631 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33ff4599352ab484361317e1141542245cd3827b024c49d1b281f4433e833f9c\": container with ID starting with 33ff4599352ab484361317e1141542245cd3827b024c49d1b281f4433e833f9c not found: ID does not exist" containerID="33ff4599352ab484361317e1141542245cd3827b024c49d1b281f4433e833f9c" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.647660 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33ff4599352ab484361317e1141542245cd3827b024c49d1b281f4433e833f9c"} err="failed to get container status \"33ff4599352ab484361317e1141542245cd3827b024c49d1b281f4433e833f9c\": rpc error: code = NotFound desc = could not find container \"33ff4599352ab484361317e1141542245cd3827b024c49d1b281f4433e833f9c\": container with ID starting with 33ff4599352ab484361317e1141542245cd3827b024c49d1b281f4433e833f9c not found: ID does not exist" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.647672 4904 scope.go:117] "RemoveContainer" containerID="62d391f394fced8cef03cbba49f457f6818005d1c190e5ca767d45bbc9d57a2c" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.658955 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-977kk"] Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.661196 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-977kk"] Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.662763 4904 scope.go:117] "RemoveContainer" containerID="866702745cde4a085d2378b9245f45e6594a426b09b695b26c815e7de28e2211" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.663654 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63c9ce1a-01e3-42f3-8027-bfda9be0b3ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63c9ce1a-01e3-42f3-8027-bfda9be0b3ff" (UID: "63c9ce1a-01e3-42f3-8027-bfda9be0b3ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.683245 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nvpkp"] Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.683583 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nvpkp"] Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.704509 4904 scope.go:117] "RemoveContainer" containerID="22e81e28a495ab6f07d67874f918993f4cc3d8586268e06e326f31c954e5e63b" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.732637 4904 scope.go:117] "RemoveContainer" containerID="62d391f394fced8cef03cbba49f457f6818005d1c190e5ca767d45bbc9d57a2c" Feb 14 11:14:01 crc kubenswrapper[4904]: E0214 11:14:01.734057 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62d391f394fced8cef03cbba49f457f6818005d1c190e5ca767d45bbc9d57a2c\": container with ID starting with 62d391f394fced8cef03cbba49f457f6818005d1c190e5ca767d45bbc9d57a2c not found: ID does not exist" containerID="62d391f394fced8cef03cbba49f457f6818005d1c190e5ca767d45bbc9d57a2c" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.734100 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62d391f394fced8cef03cbba49f457f6818005d1c190e5ca767d45bbc9d57a2c"} err="failed to get container status \"62d391f394fced8cef03cbba49f457f6818005d1c190e5ca767d45bbc9d57a2c\": rpc error: code = NotFound desc = could not find container \"62d391f394fced8cef03cbba49f457f6818005d1c190e5ca767d45bbc9d57a2c\": container with ID starting with 62d391f394fced8cef03cbba49f457f6818005d1c190e5ca767d45bbc9d57a2c not found: ID does not exist" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.734131 4904 scope.go:117] "RemoveContainer" containerID="866702745cde4a085d2378b9245f45e6594a426b09b695b26c815e7de28e2211" Feb 14 11:14:01 crc kubenswrapper[4904]: E0214 11:14:01.734736 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"866702745cde4a085d2378b9245f45e6594a426b09b695b26c815e7de28e2211\": container with ID starting with 866702745cde4a085d2378b9245f45e6594a426b09b695b26c815e7de28e2211 not found: ID does not exist" containerID="866702745cde4a085d2378b9245f45e6594a426b09b695b26c815e7de28e2211" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.734848 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"866702745cde4a085d2378b9245f45e6594a426b09b695b26c815e7de28e2211"} err="failed to get container status \"866702745cde4a085d2378b9245f45e6594a426b09b695b26c815e7de28e2211\": rpc error: code = NotFound desc = could not find container \"866702745cde4a085d2378b9245f45e6594a426b09b695b26c815e7de28e2211\": container with ID starting with 866702745cde4a085d2378b9245f45e6594a426b09b695b26c815e7de28e2211 not found: ID does not exist" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.734873 4904 scope.go:117] "RemoveContainer" containerID="22e81e28a495ab6f07d67874f918993f4cc3d8586268e06e326f31c954e5e63b" Feb 14 11:14:01 crc kubenswrapper[4904]: E0214 11:14:01.737970 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22e81e28a495ab6f07d67874f918993f4cc3d8586268e06e326f31c954e5e63b\": container with ID starting with 22e81e28a495ab6f07d67874f918993f4cc3d8586268e06e326f31c954e5e63b not found: ID does not exist" containerID="22e81e28a495ab6f07d67874f918993f4cc3d8586268e06e326f31c954e5e63b" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.738004 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22e81e28a495ab6f07d67874f918993f4cc3d8586268e06e326f31c954e5e63b"} err="failed to get container status \"22e81e28a495ab6f07d67874f918993f4cc3d8586268e06e326f31c954e5e63b\": rpc error: code = NotFound desc = could not find container \"22e81e28a495ab6f07d67874f918993f4cc3d8586268e06e326f31c954e5e63b\": container with ID starting with 22e81e28a495ab6f07d67874f918993f4cc3d8586268e06e326f31c954e5e63b not found: ID does not exist" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.738027 4904 scope.go:117] "RemoveContainer" containerID="6c243d2633d6cd00022d141ab70d97df2920850cf29e1c198bbc7684c36ac17f" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.744022 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63c9ce1a-01e3-42f3-8027-bfda9be0b3ff-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.749028 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.783448 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49e91378-b5ff-4480-abbf-765ccc061978-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49e91378-b5ff-4480-abbf-765ccc061978" (UID: "49e91378-b5ff-4480-abbf-765ccc061978"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.790496 4904 scope.go:117] "RemoveContainer" containerID="566e1b4c9fb87df37d5575f8681b1a8c1074cae72fcf186958f564348cce1b9f" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.820945 4904 scope.go:117] "RemoveContainer" containerID="3b8956f5878119fde34b8d263a61fdaa2af4aa0cefe3aabd7f94eea092a181f3" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.845212 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f130a8a5-24b3-4916-a0c7-b1fa5e55cad4-marketplace-operator-metrics\") pod \"f130a8a5-24b3-4916-a0c7-b1fa5e55cad4\" (UID: \"f130a8a5-24b3-4916-a0c7-b1fa5e55cad4\") " Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.845418 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzzg8\" (UniqueName: \"kubernetes.io/projected/f130a8a5-24b3-4916-a0c7-b1fa5e55cad4-kube-api-access-pzzg8\") pod \"f130a8a5-24b3-4916-a0c7-b1fa5e55cad4\" (UID: \"f130a8a5-24b3-4916-a0c7-b1fa5e55cad4\") " Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.845538 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f130a8a5-24b3-4916-a0c7-b1fa5e55cad4-marketplace-trusted-ca\") pod \"f130a8a5-24b3-4916-a0c7-b1fa5e55cad4\" (UID: \"f130a8a5-24b3-4916-a0c7-b1fa5e55cad4\") " Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.845815 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49e91378-b5ff-4480-abbf-765ccc061978-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.846671 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f130a8a5-24b3-4916-a0c7-b1fa5e55cad4-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "f130a8a5-24b3-4916-a0c7-b1fa5e55cad4" (UID: "f130a8a5-24b3-4916-a0c7-b1fa5e55cad4"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.857502 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f130a8a5-24b3-4916-a0c7-b1fa5e55cad4-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "f130a8a5-24b3-4916-a0c7-b1fa5e55cad4" (UID: "f130a8a5-24b3-4916-a0c7-b1fa5e55cad4"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.857626 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f130a8a5-24b3-4916-a0c7-b1fa5e55cad4-kube-api-access-pzzg8" (OuterVolumeSpecName: "kube-api-access-pzzg8") pod "f130a8a5-24b3-4916-a0c7-b1fa5e55cad4" (UID: "f130a8a5-24b3-4916-a0c7-b1fa5e55cad4"). InnerVolumeSpecName "kube-api-access-pzzg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.861015 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92dc9e32-c069-4de6-a3ae-28b80b60d722" path="/var/lib/kubelet/pods/92dc9e32-c069-4de6-a3ae-28b80b60d722/volumes" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.861892 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94b8c7dc-4ce9-4434-a387-e19139fe8a3d" path="/var/lib/kubelet/pods/94b8c7dc-4ce9-4434-a387-e19139fe8a3d/volumes" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.899050 4904 scope.go:117] "RemoveContainer" containerID="6c243d2633d6cd00022d141ab70d97df2920850cf29e1c198bbc7684c36ac17f" Feb 14 11:14:01 crc kubenswrapper[4904]: E0214 11:14:01.899822 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c243d2633d6cd00022d141ab70d97df2920850cf29e1c198bbc7684c36ac17f\": container with ID starting with 6c243d2633d6cd00022d141ab70d97df2920850cf29e1c198bbc7684c36ac17f not found: ID does not exist" containerID="6c243d2633d6cd00022d141ab70d97df2920850cf29e1c198bbc7684c36ac17f" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.899878 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c243d2633d6cd00022d141ab70d97df2920850cf29e1c198bbc7684c36ac17f"} err="failed to get container status \"6c243d2633d6cd00022d141ab70d97df2920850cf29e1c198bbc7684c36ac17f\": rpc error: code = NotFound desc = could not find container \"6c243d2633d6cd00022d141ab70d97df2920850cf29e1c198bbc7684c36ac17f\": container with ID starting with 6c243d2633d6cd00022d141ab70d97df2920850cf29e1c198bbc7684c36ac17f not found: ID does not exist" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.899923 4904 scope.go:117] "RemoveContainer" containerID="566e1b4c9fb87df37d5575f8681b1a8c1074cae72fcf186958f564348cce1b9f" Feb 14 11:14:01 crc kubenswrapper[4904]: E0214 11:14:01.902302 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"566e1b4c9fb87df37d5575f8681b1a8c1074cae72fcf186958f564348cce1b9f\": container with ID starting with 566e1b4c9fb87df37d5575f8681b1a8c1074cae72fcf186958f564348cce1b9f not found: ID does not exist" containerID="566e1b4c9fb87df37d5575f8681b1a8c1074cae72fcf186958f564348cce1b9f" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.902336 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"566e1b4c9fb87df37d5575f8681b1a8c1074cae72fcf186958f564348cce1b9f"} err="failed to get container status \"566e1b4c9fb87df37d5575f8681b1a8c1074cae72fcf186958f564348cce1b9f\": rpc error: code = NotFound desc = could not find container \"566e1b4c9fb87df37d5575f8681b1a8c1074cae72fcf186958f564348cce1b9f\": container with ID starting with 566e1b4c9fb87df37d5575f8681b1a8c1074cae72fcf186958f564348cce1b9f not found: ID does not exist" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.902362 4904 scope.go:117] "RemoveContainer" containerID="3b8956f5878119fde34b8d263a61fdaa2af4aa0cefe3aabd7f94eea092a181f3" Feb 14 11:14:01 crc kubenswrapper[4904]: E0214 11:14:01.902743 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b8956f5878119fde34b8d263a61fdaa2af4aa0cefe3aabd7f94eea092a181f3\": container with ID starting with 3b8956f5878119fde34b8d263a61fdaa2af4aa0cefe3aabd7f94eea092a181f3 not found: ID does not exist" containerID="3b8956f5878119fde34b8d263a61fdaa2af4aa0cefe3aabd7f94eea092a181f3" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.902762 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b8956f5878119fde34b8d263a61fdaa2af4aa0cefe3aabd7f94eea092a181f3"} err="failed to get container status \"3b8956f5878119fde34b8d263a61fdaa2af4aa0cefe3aabd7f94eea092a181f3\": rpc error: code = NotFound desc = could not find container \"3b8956f5878119fde34b8d263a61fdaa2af4aa0cefe3aabd7f94eea092a181f3\": container with ID starting with 3b8956f5878119fde34b8d263a61fdaa2af4aa0cefe3aabd7f94eea092a181f3 not found: ID does not exist" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.902774 4904 scope.go:117] "RemoveContainer" containerID="91278d384c32444678079c8fda4688f16a28bc0032d085b4a9216e1b1801b357" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.923587 4904 scope.go:117] "RemoveContainer" containerID="41e52ddb95a6e2f4d6c0f8a62704a328ef23392029f9c947a665d03972e53e12" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.947311 4904 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f130a8a5-24b3-4916-a0c7-b1fa5e55cad4-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.947342 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzzg8\" (UniqueName: \"kubernetes.io/projected/f130a8a5-24b3-4916-a0c7-b1fa5e55cad4-kube-api-access-pzzg8\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.947350 4904 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f130a8a5-24b3-4916-a0c7-b1fa5e55cad4-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.948385 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-krh2c"] Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.962478 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-krh2c"] Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.962512 4904 scope.go:117] "RemoveContainer" containerID="20f6bcba8893436605e59344de145196a5f5ed8e267d59dff7f682eaf65f10ec" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.966702 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bvpm5"] Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.969714 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bvpm5"] Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.976664 4904 scope.go:117] "RemoveContainer" containerID="91278d384c32444678079c8fda4688f16a28bc0032d085b4a9216e1b1801b357" Feb 14 11:14:01 crc kubenswrapper[4904]: E0214 11:14:01.977191 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91278d384c32444678079c8fda4688f16a28bc0032d085b4a9216e1b1801b357\": container with ID starting with 91278d384c32444678079c8fda4688f16a28bc0032d085b4a9216e1b1801b357 not found: ID does not exist" containerID="91278d384c32444678079c8fda4688f16a28bc0032d085b4a9216e1b1801b357" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.977228 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91278d384c32444678079c8fda4688f16a28bc0032d085b4a9216e1b1801b357"} err="failed to get container status \"91278d384c32444678079c8fda4688f16a28bc0032d085b4a9216e1b1801b357\": rpc error: code = NotFound desc = could not find container \"91278d384c32444678079c8fda4688f16a28bc0032d085b4a9216e1b1801b357\": container with ID starting with 91278d384c32444678079c8fda4688f16a28bc0032d085b4a9216e1b1801b357 not found: ID does not exist" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.977260 4904 scope.go:117] "RemoveContainer" containerID="41e52ddb95a6e2f4d6c0f8a62704a328ef23392029f9c947a665d03972e53e12" Feb 14 11:14:01 crc kubenswrapper[4904]: E0214 11:14:01.977626 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41e52ddb95a6e2f4d6c0f8a62704a328ef23392029f9c947a665d03972e53e12\": container with ID starting with 41e52ddb95a6e2f4d6c0f8a62704a328ef23392029f9c947a665d03972e53e12 not found: ID does not exist" containerID="41e52ddb95a6e2f4d6c0f8a62704a328ef23392029f9c947a665d03972e53e12" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.977728 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41e52ddb95a6e2f4d6c0f8a62704a328ef23392029f9c947a665d03972e53e12"} err="failed to get container status \"41e52ddb95a6e2f4d6c0f8a62704a328ef23392029f9c947a665d03972e53e12\": rpc error: code = NotFound desc = could not find container \"41e52ddb95a6e2f4d6c0f8a62704a328ef23392029f9c947a665d03972e53e12\": container with ID starting with 41e52ddb95a6e2f4d6c0f8a62704a328ef23392029f9c947a665d03972e53e12 not found: ID does not exist" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.977816 4904 scope.go:117] "RemoveContainer" containerID="20f6bcba8893436605e59344de145196a5f5ed8e267d59dff7f682eaf65f10ec" Feb 14 11:14:01 crc kubenswrapper[4904]: E0214 11:14:01.978236 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20f6bcba8893436605e59344de145196a5f5ed8e267d59dff7f682eaf65f10ec\": container with ID starting with 20f6bcba8893436605e59344de145196a5f5ed8e267d59dff7f682eaf65f10ec not found: ID does not exist" containerID="20f6bcba8893436605e59344de145196a5f5ed8e267d59dff7f682eaf65f10ec" Feb 14 11:14:01 crc kubenswrapper[4904]: I0214 11:14:01.978260 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20f6bcba8893436605e59344de145196a5f5ed8e267d59dff7f682eaf65f10ec"} err="failed to get container status \"20f6bcba8893436605e59344de145196a5f5ed8e267d59dff7f682eaf65f10ec\": rpc error: code = NotFound desc = could not find container \"20f6bcba8893436605e59344de145196a5f5ed8e267d59dff7f682eaf65f10ec\": container with ID starting with 20f6bcba8893436605e59344de145196a5f5ed8e267d59dff7f682eaf65f10ec not found: ID does not exist" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.634074 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" event={"ID":"f130a8a5-24b3-4916-a0c7-b1fa5e55cad4","Type":"ContainerDied","Data":"d3b7f6adb60da3cbff5abdf8689b57ff9d0e31b8d04cbcb123d186c35c73c7a4"} Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.634374 4904 scope.go:117] "RemoveContainer" containerID="c58cdbc463ae1de3ed0c35e320580faeaddb6a95577c99c69fe14ae19e9381c9" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.634132 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9pzkg" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.639723 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-h5v26" event={"ID":"2104a707-8d9b-43e6-9c74-f41bdafcf818","Type":"ContainerStarted","Data":"ab97db421fc062e7be5aa49871f47527e1fe1a64e0b903b357e933d13d21de1f"} Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.639782 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-h5v26" event={"ID":"2104a707-8d9b-43e6-9c74-f41bdafcf818","Type":"ContainerStarted","Data":"357690f5886d9898047eb3e070ae924699bbb67b42c2e8e14b698f22973899d7"} Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.640341 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-h5v26" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.643345 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-h5v26" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.655069 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9pzkg"] Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.659364 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9pzkg"] Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.728974 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-h5v26" podStartSLOduration=2.728953033 podStartE2EDuration="2.728953033s" podCreationTimestamp="2026-02-14 11:14:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:14:02.671802707 +0000 UTC m=+233.484567378" watchObservedRunningTime="2026-02-14 11:14:02.728953033 +0000 UTC m=+233.541717694" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.858139 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h2tbz"] Feb 14 11:14:02 crc kubenswrapper[4904]: E0214 11:14:02.858344 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f130a8a5-24b3-4916-a0c7-b1fa5e55cad4" containerName="marketplace-operator" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.858356 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="f130a8a5-24b3-4916-a0c7-b1fa5e55cad4" containerName="marketplace-operator" Feb 14 11:14:02 crc kubenswrapper[4904]: E0214 11:14:02.858365 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94b8c7dc-4ce9-4434-a387-e19139fe8a3d" containerName="registry-server" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.858372 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="94b8c7dc-4ce9-4434-a387-e19139fe8a3d" containerName="registry-server" Feb 14 11:14:02 crc kubenswrapper[4904]: E0214 11:14:02.858383 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92dc9e32-c069-4de6-a3ae-28b80b60d722" containerName="extract-content" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.858390 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="92dc9e32-c069-4de6-a3ae-28b80b60d722" containerName="extract-content" Feb 14 11:14:02 crc kubenswrapper[4904]: E0214 11:14:02.858399 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92dc9e32-c069-4de6-a3ae-28b80b60d722" containerName="extract-utilities" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.858404 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="92dc9e32-c069-4de6-a3ae-28b80b60d722" containerName="extract-utilities" Feb 14 11:14:02 crc kubenswrapper[4904]: E0214 11:14:02.858413 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e91378-b5ff-4480-abbf-765ccc061978" containerName="extract-content" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.858546 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e91378-b5ff-4480-abbf-765ccc061978" containerName="extract-content" Feb 14 11:14:02 crc kubenswrapper[4904]: E0214 11:14:02.858561 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63c9ce1a-01e3-42f3-8027-bfda9be0b3ff" containerName="registry-server" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.858567 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="63c9ce1a-01e3-42f3-8027-bfda9be0b3ff" containerName="registry-server" Feb 14 11:14:02 crc kubenswrapper[4904]: E0214 11:14:02.858580 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92dc9e32-c069-4de6-a3ae-28b80b60d722" containerName="registry-server" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.858586 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="92dc9e32-c069-4de6-a3ae-28b80b60d722" containerName="registry-server" Feb 14 11:14:02 crc kubenswrapper[4904]: E0214 11:14:02.858594 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63c9ce1a-01e3-42f3-8027-bfda9be0b3ff" containerName="extract-utilities" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.858600 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="63c9ce1a-01e3-42f3-8027-bfda9be0b3ff" containerName="extract-utilities" Feb 14 11:14:02 crc kubenswrapper[4904]: E0214 11:14:02.858606 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63c9ce1a-01e3-42f3-8027-bfda9be0b3ff" containerName="extract-content" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.858612 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="63c9ce1a-01e3-42f3-8027-bfda9be0b3ff" containerName="extract-content" Feb 14 11:14:02 crc kubenswrapper[4904]: E0214 11:14:02.858621 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94b8c7dc-4ce9-4434-a387-e19139fe8a3d" containerName="extract-content" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.858627 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="94b8c7dc-4ce9-4434-a387-e19139fe8a3d" containerName="extract-content" Feb 14 11:14:02 crc kubenswrapper[4904]: E0214 11:14:02.858634 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e91378-b5ff-4480-abbf-765ccc061978" containerName="registry-server" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.858639 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e91378-b5ff-4480-abbf-765ccc061978" containerName="registry-server" Feb 14 11:14:02 crc kubenswrapper[4904]: E0214 11:14:02.858648 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94b8c7dc-4ce9-4434-a387-e19139fe8a3d" containerName="extract-utilities" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.858654 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="94b8c7dc-4ce9-4434-a387-e19139fe8a3d" containerName="extract-utilities" Feb 14 11:14:02 crc kubenswrapper[4904]: E0214 11:14:02.858661 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e91378-b5ff-4480-abbf-765ccc061978" containerName="extract-utilities" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.858669 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e91378-b5ff-4480-abbf-765ccc061978" containerName="extract-utilities" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.858749 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="49e91378-b5ff-4480-abbf-765ccc061978" containerName="registry-server" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.858758 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="63c9ce1a-01e3-42f3-8027-bfda9be0b3ff" containerName="registry-server" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.858771 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="f130a8a5-24b3-4916-a0c7-b1fa5e55cad4" containerName="marketplace-operator" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.858778 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="94b8c7dc-4ce9-4434-a387-e19139fe8a3d" containerName="registry-server" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.858786 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="92dc9e32-c069-4de6-a3ae-28b80b60d722" containerName="registry-server" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.859641 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h2tbz" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.864646 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.867597 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h2tbz"] Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.958452 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9pxx\" (UniqueName: \"kubernetes.io/projected/9a1b8061-2687-4083-8e86-6c9b99daf8a5-kube-api-access-q9pxx\") pod \"redhat-marketplace-h2tbz\" (UID: \"9a1b8061-2687-4083-8e86-6c9b99daf8a5\") " pod="openshift-marketplace/redhat-marketplace-h2tbz" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.958524 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a1b8061-2687-4083-8e86-6c9b99daf8a5-catalog-content\") pod \"redhat-marketplace-h2tbz\" (UID: \"9a1b8061-2687-4083-8e86-6c9b99daf8a5\") " pod="openshift-marketplace/redhat-marketplace-h2tbz" Feb 14 11:14:02 crc kubenswrapper[4904]: I0214 11:14:02.958561 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a1b8061-2687-4083-8e86-6c9b99daf8a5-utilities\") pod \"redhat-marketplace-h2tbz\" (UID: \"9a1b8061-2687-4083-8e86-6c9b99daf8a5\") " pod="openshift-marketplace/redhat-marketplace-h2tbz" Feb 14 11:14:03 crc kubenswrapper[4904]: I0214 11:14:03.060071 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a1b8061-2687-4083-8e86-6c9b99daf8a5-utilities\") pod \"redhat-marketplace-h2tbz\" (UID: \"9a1b8061-2687-4083-8e86-6c9b99daf8a5\") " pod="openshift-marketplace/redhat-marketplace-h2tbz" Feb 14 11:14:03 crc kubenswrapper[4904]: I0214 11:14:03.060152 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9pxx\" (UniqueName: \"kubernetes.io/projected/9a1b8061-2687-4083-8e86-6c9b99daf8a5-kube-api-access-q9pxx\") pod \"redhat-marketplace-h2tbz\" (UID: \"9a1b8061-2687-4083-8e86-6c9b99daf8a5\") " pod="openshift-marketplace/redhat-marketplace-h2tbz" Feb 14 11:14:03 crc kubenswrapper[4904]: I0214 11:14:03.060218 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a1b8061-2687-4083-8e86-6c9b99daf8a5-catalog-content\") pod \"redhat-marketplace-h2tbz\" (UID: \"9a1b8061-2687-4083-8e86-6c9b99daf8a5\") " pod="openshift-marketplace/redhat-marketplace-h2tbz" Feb 14 11:14:03 crc kubenswrapper[4904]: I0214 11:14:03.060526 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a1b8061-2687-4083-8e86-6c9b99daf8a5-utilities\") pod \"redhat-marketplace-h2tbz\" (UID: \"9a1b8061-2687-4083-8e86-6c9b99daf8a5\") " pod="openshift-marketplace/redhat-marketplace-h2tbz" Feb 14 11:14:03 crc kubenswrapper[4904]: I0214 11:14:03.060886 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a1b8061-2687-4083-8e86-6c9b99daf8a5-catalog-content\") pod \"redhat-marketplace-h2tbz\" (UID: \"9a1b8061-2687-4083-8e86-6c9b99daf8a5\") " pod="openshift-marketplace/redhat-marketplace-h2tbz" Feb 14 11:14:03 crc kubenswrapper[4904]: I0214 11:14:03.085061 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9pxx\" (UniqueName: \"kubernetes.io/projected/9a1b8061-2687-4083-8e86-6c9b99daf8a5-kube-api-access-q9pxx\") pod \"redhat-marketplace-h2tbz\" (UID: \"9a1b8061-2687-4083-8e86-6c9b99daf8a5\") " pod="openshift-marketplace/redhat-marketplace-h2tbz" Feb 14 11:14:03 crc kubenswrapper[4904]: I0214 11:14:03.190505 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h2tbz" Feb 14 11:14:03 crc kubenswrapper[4904]: I0214 11:14:03.591102 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h2tbz"] Feb 14 11:14:03 crc kubenswrapper[4904]: I0214 11:14:03.652252 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2tbz" event={"ID":"9a1b8061-2687-4083-8e86-6c9b99daf8a5","Type":"ContainerStarted","Data":"c683de05a8d8332dd5e6182e2932ba3efc4581b588191f95b6ce80416936a68f"} Feb 14 11:14:03 crc kubenswrapper[4904]: I0214 11:14:03.846066 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49e91378-b5ff-4480-abbf-765ccc061978" path="/var/lib/kubelet/pods/49e91378-b5ff-4480-abbf-765ccc061978/volumes" Feb 14 11:14:03 crc kubenswrapper[4904]: I0214 11:14:03.847216 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63c9ce1a-01e3-42f3-8027-bfda9be0b3ff" path="/var/lib/kubelet/pods/63c9ce1a-01e3-42f3-8027-bfda9be0b3ff/volumes" Feb 14 11:14:03 crc kubenswrapper[4904]: I0214 11:14:03.848274 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f130a8a5-24b3-4916-a0c7-b1fa5e55cad4" path="/var/lib/kubelet/pods/f130a8a5-24b3-4916-a0c7-b1fa5e55cad4/volumes" Feb 14 11:14:04 crc kubenswrapper[4904]: I0214 11:14:04.255758 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-99bfv"] Feb 14 11:14:04 crc kubenswrapper[4904]: I0214 11:14:04.257046 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-99bfv" Feb 14 11:14:04 crc kubenswrapper[4904]: I0214 11:14:04.258869 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 14 11:14:04 crc kubenswrapper[4904]: I0214 11:14:04.273421 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-99bfv"] Feb 14 11:14:04 crc kubenswrapper[4904]: I0214 11:14:04.378455 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4233cb32-d700-4c17-a97b-82d5bcd6ea5f-catalog-content\") pod \"community-operators-99bfv\" (UID: \"4233cb32-d700-4c17-a97b-82d5bcd6ea5f\") " pod="openshift-marketplace/community-operators-99bfv" Feb 14 11:14:04 crc kubenswrapper[4904]: I0214 11:14:04.378514 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4233cb32-d700-4c17-a97b-82d5bcd6ea5f-utilities\") pod \"community-operators-99bfv\" (UID: \"4233cb32-d700-4c17-a97b-82d5bcd6ea5f\") " pod="openshift-marketplace/community-operators-99bfv" Feb 14 11:14:04 crc kubenswrapper[4904]: I0214 11:14:04.378543 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jrpl\" (UniqueName: \"kubernetes.io/projected/4233cb32-d700-4c17-a97b-82d5bcd6ea5f-kube-api-access-7jrpl\") pod \"community-operators-99bfv\" (UID: \"4233cb32-d700-4c17-a97b-82d5bcd6ea5f\") " pod="openshift-marketplace/community-operators-99bfv" Feb 14 11:14:04 crc kubenswrapper[4904]: I0214 11:14:04.479308 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4233cb32-d700-4c17-a97b-82d5bcd6ea5f-catalog-content\") pod \"community-operators-99bfv\" (UID: \"4233cb32-d700-4c17-a97b-82d5bcd6ea5f\") " pod="openshift-marketplace/community-operators-99bfv" Feb 14 11:14:04 crc kubenswrapper[4904]: I0214 11:14:04.479353 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4233cb32-d700-4c17-a97b-82d5bcd6ea5f-utilities\") pod \"community-operators-99bfv\" (UID: \"4233cb32-d700-4c17-a97b-82d5bcd6ea5f\") " pod="openshift-marketplace/community-operators-99bfv" Feb 14 11:14:04 crc kubenswrapper[4904]: I0214 11:14:04.479386 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jrpl\" (UniqueName: \"kubernetes.io/projected/4233cb32-d700-4c17-a97b-82d5bcd6ea5f-kube-api-access-7jrpl\") pod \"community-operators-99bfv\" (UID: \"4233cb32-d700-4c17-a97b-82d5bcd6ea5f\") " pod="openshift-marketplace/community-operators-99bfv" Feb 14 11:14:04 crc kubenswrapper[4904]: I0214 11:14:04.479969 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4233cb32-d700-4c17-a97b-82d5bcd6ea5f-utilities\") pod \"community-operators-99bfv\" (UID: \"4233cb32-d700-4c17-a97b-82d5bcd6ea5f\") " pod="openshift-marketplace/community-operators-99bfv" Feb 14 11:14:04 crc kubenswrapper[4904]: I0214 11:14:04.480052 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4233cb32-d700-4c17-a97b-82d5bcd6ea5f-catalog-content\") pod \"community-operators-99bfv\" (UID: \"4233cb32-d700-4c17-a97b-82d5bcd6ea5f\") " pod="openshift-marketplace/community-operators-99bfv" Feb 14 11:14:04 crc kubenswrapper[4904]: I0214 11:14:04.498810 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jrpl\" (UniqueName: \"kubernetes.io/projected/4233cb32-d700-4c17-a97b-82d5bcd6ea5f-kube-api-access-7jrpl\") pod \"community-operators-99bfv\" (UID: \"4233cb32-d700-4c17-a97b-82d5bcd6ea5f\") " pod="openshift-marketplace/community-operators-99bfv" Feb 14 11:14:04 crc kubenswrapper[4904]: I0214 11:14:04.576755 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-99bfv" Feb 14 11:14:04 crc kubenswrapper[4904]: I0214 11:14:04.664395 4904 generic.go:334] "Generic (PLEG): container finished" podID="9a1b8061-2687-4083-8e86-6c9b99daf8a5" containerID="564bc67b2b17b2af313184029c98eea21d9253c2ea73287f092e77c9005c5dfc" exitCode=0 Feb 14 11:14:04 crc kubenswrapper[4904]: I0214 11:14:04.664466 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2tbz" event={"ID":"9a1b8061-2687-4083-8e86-6c9b99daf8a5","Type":"ContainerDied","Data":"564bc67b2b17b2af313184029c98eea21d9253c2ea73287f092e77c9005c5dfc"} Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.033858 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-99bfv"] Feb 14 11:14:05 crc kubenswrapper[4904]: W0214 11:14:05.039831 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4233cb32_d700_4c17_a97b_82d5bcd6ea5f.slice/crio-5222942d7612d39ae22140c5b5465ddfaa271272bc6726f38a059867ba404eb9 WatchSource:0}: Error finding container 5222942d7612d39ae22140c5b5465ddfaa271272bc6726f38a059867ba404eb9: Status 404 returned error can't find the container with id 5222942d7612d39ae22140c5b5465ddfaa271272bc6726f38a059867ba404eb9 Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.265716 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4m6zf"] Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.267131 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4m6zf" Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.269308 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.271744 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4m6zf"] Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.412775 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfb9b35f-85bb-43ec-a29d-897b3ebf0da9-catalog-content\") pod \"certified-operators-4m6zf\" (UID: \"dfb9b35f-85bb-43ec-a29d-897b3ebf0da9\") " pod="openshift-marketplace/certified-operators-4m6zf" Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.412812 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfb9b35f-85bb-43ec-a29d-897b3ebf0da9-utilities\") pod \"certified-operators-4m6zf\" (UID: \"dfb9b35f-85bb-43ec-a29d-897b3ebf0da9\") " pod="openshift-marketplace/certified-operators-4m6zf" Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.412861 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcq6j\" (UniqueName: \"kubernetes.io/projected/dfb9b35f-85bb-43ec-a29d-897b3ebf0da9-kube-api-access-zcq6j\") pod \"certified-operators-4m6zf\" (UID: \"dfb9b35f-85bb-43ec-a29d-897b3ebf0da9\") " pod="openshift-marketplace/certified-operators-4m6zf" Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.514419 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfb9b35f-85bb-43ec-a29d-897b3ebf0da9-utilities\") pod \"certified-operators-4m6zf\" (UID: \"dfb9b35f-85bb-43ec-a29d-897b3ebf0da9\") " pod="openshift-marketplace/certified-operators-4m6zf" Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.514462 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfb9b35f-85bb-43ec-a29d-897b3ebf0da9-catalog-content\") pod \"certified-operators-4m6zf\" (UID: \"dfb9b35f-85bb-43ec-a29d-897b3ebf0da9\") " pod="openshift-marketplace/certified-operators-4m6zf" Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.514496 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcq6j\" (UniqueName: \"kubernetes.io/projected/dfb9b35f-85bb-43ec-a29d-897b3ebf0da9-kube-api-access-zcq6j\") pod \"certified-operators-4m6zf\" (UID: \"dfb9b35f-85bb-43ec-a29d-897b3ebf0da9\") " pod="openshift-marketplace/certified-operators-4m6zf" Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.515153 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfb9b35f-85bb-43ec-a29d-897b3ebf0da9-utilities\") pod \"certified-operators-4m6zf\" (UID: \"dfb9b35f-85bb-43ec-a29d-897b3ebf0da9\") " pod="openshift-marketplace/certified-operators-4m6zf" Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.515195 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfb9b35f-85bb-43ec-a29d-897b3ebf0da9-catalog-content\") pod \"certified-operators-4m6zf\" (UID: \"dfb9b35f-85bb-43ec-a29d-897b3ebf0da9\") " pod="openshift-marketplace/certified-operators-4m6zf" Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.532115 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcq6j\" (UniqueName: \"kubernetes.io/projected/dfb9b35f-85bb-43ec-a29d-897b3ebf0da9-kube-api-access-zcq6j\") pod \"certified-operators-4m6zf\" (UID: \"dfb9b35f-85bb-43ec-a29d-897b3ebf0da9\") " pod="openshift-marketplace/certified-operators-4m6zf" Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.583861 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4m6zf" Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.672062 4904 generic.go:334] "Generic (PLEG): container finished" podID="4233cb32-d700-4c17-a97b-82d5bcd6ea5f" containerID="84742fc25cb913c30e2bae476217fd4bb581fa83787afb7d746b0a0ab2e34d5e" exitCode=0 Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.672142 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-99bfv" event={"ID":"4233cb32-d700-4c17-a97b-82d5bcd6ea5f","Type":"ContainerDied","Data":"84742fc25cb913c30e2bae476217fd4bb581fa83787afb7d746b0a0ab2e34d5e"} Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.672180 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-99bfv" event={"ID":"4233cb32-d700-4c17-a97b-82d5bcd6ea5f","Type":"ContainerStarted","Data":"5222942d7612d39ae22140c5b5465ddfaa271272bc6726f38a059867ba404eb9"} Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.683804 4904 generic.go:334] "Generic (PLEG): container finished" podID="9a1b8061-2687-4083-8e86-6c9b99daf8a5" containerID="0c27e60f7f40a379049a3f84e55e660b0f6ff706766a24f6d2efdb155a9b30b3" exitCode=0 Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.683882 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2tbz" event={"ID":"9a1b8061-2687-4083-8e86-6c9b99daf8a5","Type":"ContainerDied","Data":"0c27e60f7f40a379049a3f84e55e660b0f6ff706766a24f6d2efdb155a9b30b3"} Feb 14 11:14:05 crc kubenswrapper[4904]: I0214 11:14:05.963515 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4m6zf"] Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.657201 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wxxxn"] Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.658426 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wxxxn" Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.664169 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.674266 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wxxxn"] Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.699427 4904 generic.go:334] "Generic (PLEG): container finished" podID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" containerID="c684deaf5bae729345458edea8de10a6aebc187fd713b5a8c133ea742d1468b4" exitCode=0 Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.699480 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4m6zf" event={"ID":"dfb9b35f-85bb-43ec-a29d-897b3ebf0da9","Type":"ContainerDied","Data":"c684deaf5bae729345458edea8de10a6aebc187fd713b5a8c133ea742d1468b4"} Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.699540 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4m6zf" event={"ID":"dfb9b35f-85bb-43ec-a29d-897b3ebf0da9","Type":"ContainerStarted","Data":"85b20507c3fff49c6333a44f82228a8d7531af37795a888ba8480576064f8818"} Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.701393 4904 generic.go:334] "Generic (PLEG): container finished" podID="4233cb32-d700-4c17-a97b-82d5bcd6ea5f" containerID="b2ee2cb01541930265cedccd32de35952eb2752eb47ed2095ca4ffae5d714b03" exitCode=0 Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.702122 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-99bfv" event={"ID":"4233cb32-d700-4c17-a97b-82d5bcd6ea5f","Type":"ContainerDied","Data":"b2ee2cb01541930265cedccd32de35952eb2752eb47ed2095ca4ffae5d714b03"} Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.704296 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2tbz" event={"ID":"9a1b8061-2687-4083-8e86-6c9b99daf8a5","Type":"ContainerStarted","Data":"f9546498219da051d00d0206b3ef4e811710f309754619a2575c91ef837eb7db"} Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.770009 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h2tbz" podStartSLOduration=3.3190215739999998 podStartE2EDuration="4.769988508s" podCreationTimestamp="2026-02-14 11:14:02 +0000 UTC" firstStartedPulling="2026-02-14 11:14:04.666498988 +0000 UTC m=+235.479263649" lastFinishedPulling="2026-02-14 11:14:06.117465922 +0000 UTC m=+236.930230583" observedRunningTime="2026-02-14 11:14:06.765694018 +0000 UTC m=+237.578458679" watchObservedRunningTime="2026-02-14 11:14:06.769988508 +0000 UTC m=+237.582753179" Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.830571 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djj8l\" (UniqueName: \"kubernetes.io/projected/6b349538-4188-447c-aee9-74797e324398-kube-api-access-djj8l\") pod \"redhat-operators-wxxxn\" (UID: \"6b349538-4188-447c-aee9-74797e324398\") " pod="openshift-marketplace/redhat-operators-wxxxn" Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.830665 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b349538-4188-447c-aee9-74797e324398-catalog-content\") pod \"redhat-operators-wxxxn\" (UID: \"6b349538-4188-447c-aee9-74797e324398\") " pod="openshift-marketplace/redhat-operators-wxxxn" Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.830740 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b349538-4188-447c-aee9-74797e324398-utilities\") pod \"redhat-operators-wxxxn\" (UID: \"6b349538-4188-447c-aee9-74797e324398\") " pod="openshift-marketplace/redhat-operators-wxxxn" Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.932471 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djj8l\" (UniqueName: \"kubernetes.io/projected/6b349538-4188-447c-aee9-74797e324398-kube-api-access-djj8l\") pod \"redhat-operators-wxxxn\" (UID: \"6b349538-4188-447c-aee9-74797e324398\") " pod="openshift-marketplace/redhat-operators-wxxxn" Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.932561 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b349538-4188-447c-aee9-74797e324398-catalog-content\") pod \"redhat-operators-wxxxn\" (UID: \"6b349538-4188-447c-aee9-74797e324398\") " pod="openshift-marketplace/redhat-operators-wxxxn" Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.932586 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b349538-4188-447c-aee9-74797e324398-utilities\") pod \"redhat-operators-wxxxn\" (UID: \"6b349538-4188-447c-aee9-74797e324398\") " pod="openshift-marketplace/redhat-operators-wxxxn" Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.933327 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b349538-4188-447c-aee9-74797e324398-utilities\") pod \"redhat-operators-wxxxn\" (UID: \"6b349538-4188-447c-aee9-74797e324398\") " pod="openshift-marketplace/redhat-operators-wxxxn" Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.933524 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b349538-4188-447c-aee9-74797e324398-catalog-content\") pod \"redhat-operators-wxxxn\" (UID: \"6b349538-4188-447c-aee9-74797e324398\") " pod="openshift-marketplace/redhat-operators-wxxxn" Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.953975 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djj8l\" (UniqueName: \"kubernetes.io/projected/6b349538-4188-447c-aee9-74797e324398-kube-api-access-djj8l\") pod \"redhat-operators-wxxxn\" (UID: \"6b349538-4188-447c-aee9-74797e324398\") " pod="openshift-marketplace/redhat-operators-wxxxn" Feb 14 11:14:06 crc kubenswrapper[4904]: I0214 11:14:06.974201 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wxxxn" Feb 14 11:14:07 crc kubenswrapper[4904]: I0214 11:14:07.389087 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wxxxn"] Feb 14 11:14:07 crc kubenswrapper[4904]: W0214 11:14:07.396024 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b349538_4188_447c_aee9_74797e324398.slice/crio-7bc0d1e0a2676a650dcc42ea09c5c8580c6896ed1dabf2f849ae12f51e8d86a6 WatchSource:0}: Error finding container 7bc0d1e0a2676a650dcc42ea09c5c8580c6896ed1dabf2f849ae12f51e8d86a6: Status 404 returned error can't find the container with id 7bc0d1e0a2676a650dcc42ea09c5c8580c6896ed1dabf2f849ae12f51e8d86a6 Feb 14 11:14:07 crc kubenswrapper[4904]: I0214 11:14:07.709617 4904 generic.go:334] "Generic (PLEG): container finished" podID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" containerID="7de2cf2a14e0528c3f5b76072f34860dd8bace169bdbf1f7687044703a186b64" exitCode=0 Feb 14 11:14:07 crc kubenswrapper[4904]: I0214 11:14:07.709685 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4m6zf" event={"ID":"dfb9b35f-85bb-43ec-a29d-897b3ebf0da9","Type":"ContainerDied","Data":"7de2cf2a14e0528c3f5b76072f34860dd8bace169bdbf1f7687044703a186b64"} Feb 14 11:14:07 crc kubenswrapper[4904]: I0214 11:14:07.713809 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-99bfv" event={"ID":"4233cb32-d700-4c17-a97b-82d5bcd6ea5f","Type":"ContainerStarted","Data":"f8977c22f9afb55525c59ccf30548f4a6580c3e5ff9f4cb988d3682d55e9a8eb"} Feb 14 11:14:07 crc kubenswrapper[4904]: I0214 11:14:07.715824 4904 generic.go:334] "Generic (PLEG): container finished" podID="6b349538-4188-447c-aee9-74797e324398" containerID="382d6950833257e22469852e0f0cb02f934761d46d84385b5476092e0ba1897b" exitCode=0 Feb 14 11:14:07 crc kubenswrapper[4904]: I0214 11:14:07.715875 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wxxxn" event={"ID":"6b349538-4188-447c-aee9-74797e324398","Type":"ContainerDied","Data":"382d6950833257e22469852e0f0cb02f934761d46d84385b5476092e0ba1897b"} Feb 14 11:14:07 crc kubenswrapper[4904]: I0214 11:14:07.715909 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wxxxn" event={"ID":"6b349538-4188-447c-aee9-74797e324398","Type":"ContainerStarted","Data":"7bc0d1e0a2676a650dcc42ea09c5c8580c6896ed1dabf2f849ae12f51e8d86a6"} Feb 14 11:14:07 crc kubenswrapper[4904]: I0214 11:14:07.751352 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-99bfv" podStartSLOduration=2.316249297 podStartE2EDuration="3.751338994s" podCreationTimestamp="2026-02-14 11:14:04 +0000 UTC" firstStartedPulling="2026-02-14 11:14:05.67399704 +0000 UTC m=+236.486761691" lastFinishedPulling="2026-02-14 11:14:07.109086727 +0000 UTC m=+237.921851388" observedRunningTime="2026-02-14 11:14:07.749168413 +0000 UTC m=+238.561933094" watchObservedRunningTime="2026-02-14 11:14:07.751338994 +0000 UTC m=+238.564103655" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.095961 4904 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.096723 4904 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.096757 4904 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 14 11:14:08 crc kubenswrapper[4904]: E0214 11:14:08.096883 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.096893 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 14 11:14:08 crc kubenswrapper[4904]: E0214 11:14:08.096901 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.096907 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 14 11:14:08 crc kubenswrapper[4904]: E0214 11:14:08.096914 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.096919 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 14 11:14:08 crc kubenswrapper[4904]: E0214 11:14:08.096928 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.096933 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 14 11:14:08 crc kubenswrapper[4904]: E0214 11:14:08.096942 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.096947 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 14 11:14:08 crc kubenswrapper[4904]: E0214 11:14:08.096956 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.096961 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.097035 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.097047 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.097053 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.097060 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.097070 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.097076 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 14 11:14:08 crc kubenswrapper[4904]: E0214 11:14:08.097149 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.097158 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.099853 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634" gracePeriod=15 Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.100005 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90" gracePeriod=15 Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.100067 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37" gracePeriod=15 Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.100189 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f" gracePeriod=15 Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.100350 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e" gracePeriod=15 Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.109511 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.123815 4904 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Feb 14 11:14:08 crc kubenswrapper[4904]: E0214 11:14:08.180123 4904 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.17:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.248415 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.248472 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.248489 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.248517 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.248534 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.248551 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.248585 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.248605 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: E0214 11:14:08.269866 4904 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.17:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-operators-wxxxn.189418a4804152bc openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-operators-wxxxn,UID:6b349538-4188-447c-aee9-74797e324398,APIVersion:v1,ResourceVersion:29663,FieldPath:spec.initContainers{extract-content},},Reason:Pulled,Message:Successfully pulled image \"registry.redhat.io/redhat/redhat-operator-index:v4.18\" in 551ms (551ms including waiting). Image size: 1701129928 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-14 11:14:08.268915388 +0000 UTC m=+239.081680049,LastTimestamp:2026-02-14 11:14:08.268915388 +0000 UTC m=+239.081680049,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.349513 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.349584 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.349604 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.349637 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.349664 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.349680 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.349708 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.349726 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.349796 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.349872 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.349898 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.349922 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.349940 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.349961 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.349983 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.350002 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.480989 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.733289 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a0c8c58469c69c53355693a585add738ab568a1b126ecf179850dbd7557617b6"} Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.749757 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4m6zf" event={"ID":"dfb9b35f-85bb-43ec-a29d-897b3ebf0da9","Type":"ContainerStarted","Data":"4a0f86ee095d51e49f3054e447311afa211e612e15ba8166b6e3e28cea1254e3"} Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.750513 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.753405 4904 generic.go:334] "Generic (PLEG): container finished" podID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" containerID="1b9061fcc31f6d7899b85f1e9b619bb1363000d702e443f75da43fab2b7fb08e" exitCode=0 Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.753591 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96","Type":"ContainerDied","Data":"1b9061fcc31f6d7899b85f1e9b619bb1363000d702e443f75da43fab2b7fb08e"} Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.754307 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.754713 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.761289 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.762652 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.763481 4904 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f" exitCode=0 Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.763536 4904 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90" exitCode=0 Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.763545 4904 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e" exitCode=2 Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.763615 4904 scope.go:117] "RemoveContainer" containerID="e90dee58a3c61d06456b4396c14a81c9cb219c9558ad3d94bab6d5a6de817f75" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.767551 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wxxxn" event={"ID":"6b349538-4188-447c-aee9-74797e324398","Type":"ContainerStarted","Data":"c7ba3a8410d5f37594c5b9143892b852c22bc2d2d07aae180794b81bb2936076"} Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.768299 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.768878 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:08 crc kubenswrapper[4904]: I0214 11:14:08.769365 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:09 crc kubenswrapper[4904]: E0214 11:14:09.375919 4904 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:09 crc kubenswrapper[4904]: E0214 11:14:09.376588 4904 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:09 crc kubenswrapper[4904]: E0214 11:14:09.376768 4904 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:09 crc kubenswrapper[4904]: E0214 11:14:09.376940 4904 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:09 crc kubenswrapper[4904]: E0214 11:14:09.377105 4904 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:09 crc kubenswrapper[4904]: I0214 11:14:09.377129 4904 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 14 11:14:09 crc kubenswrapper[4904]: E0214 11:14:09.377266 4904 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="200ms" Feb 14 11:14:09 crc kubenswrapper[4904]: E0214 11:14:09.579372 4904 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="400ms" Feb 14 11:14:09 crc kubenswrapper[4904]: I0214 11:14:09.775872 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 14 11:14:09 crc kubenswrapper[4904]: I0214 11:14:09.779849 4904 generic.go:334] "Generic (PLEG): container finished" podID="6b349538-4188-447c-aee9-74797e324398" containerID="c7ba3a8410d5f37594c5b9143892b852c22bc2d2d07aae180794b81bb2936076" exitCode=0 Feb 14 11:14:09 crc kubenswrapper[4904]: I0214 11:14:09.779908 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wxxxn" event={"ID":"6b349538-4188-447c-aee9-74797e324398","Type":"ContainerDied","Data":"c7ba3a8410d5f37594c5b9143892b852c22bc2d2d07aae180794b81bb2936076"} Feb 14 11:14:09 crc kubenswrapper[4904]: I0214 11:14:09.780733 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:09 crc kubenswrapper[4904]: I0214 11:14:09.780981 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:09 crc kubenswrapper[4904]: I0214 11:14:09.781140 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:09 crc kubenswrapper[4904]: I0214 11:14:09.783695 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"11c735ef5b3fc38cf3d8214a4c0626433da291ef89e18eec6404498f2bea2c4c"} Feb 14 11:14:09 crc kubenswrapper[4904]: I0214 11:14:09.784607 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:09 crc kubenswrapper[4904]: I0214 11:14:09.784964 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:09 crc kubenswrapper[4904]: E0214 11:14:09.786047 4904 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.17:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:09 crc kubenswrapper[4904]: I0214 11:14:09.792965 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:09 crc kubenswrapper[4904]: I0214 11:14:09.840029 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:09 crc kubenswrapper[4904]: I0214 11:14:09.840493 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:09 crc kubenswrapper[4904]: I0214 11:14:09.840819 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:09 crc kubenswrapper[4904]: E0214 11:14:09.868015 4904 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.17:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-operators-wxxxn.189418a4804152bc openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-operators-wxxxn,UID:6b349538-4188-447c-aee9-74797e324398,APIVersion:v1,ResourceVersion:29663,FieldPath:spec.initContainers{extract-content},},Reason:Pulled,Message:Successfully pulled image \"registry.redhat.io/redhat/redhat-operator-index:v4.18\" in 551ms (551ms including waiting). Image size: 1701129928 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-14 11:14:08.268915388 +0000 UTC m=+239.081680049,LastTimestamp:2026-02-14 11:14:08.268915388 +0000 UTC m=+239.081680049,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 14 11:14:09 crc kubenswrapper[4904]: E0214 11:14:09.980683 4904 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="800ms" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.032374 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.032810 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.032998 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.033162 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.181015 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96-kubelet-dir\") pod \"b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96\" (UID: \"b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96\") " Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.182036 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96-var-lock\") pod \"b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96\" (UID: \"b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96\") " Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.182137 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96-kube-api-access\") pod \"b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96\" (UID: \"b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96\") " Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.182983 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" (UID: "b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.184232 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96-var-lock" (OuterVolumeSpecName: "var-lock") pod "b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" (UID: "b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.193409 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" (UID: "b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.283386 4904 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.283675 4904 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96-var-lock\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.283683 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:10 crc kubenswrapper[4904]: E0214 11:14:10.782124 4904 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="1.6s" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.794614 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wxxxn" event={"ID":"6b349538-4188-447c-aee9-74797e324398","Type":"ContainerStarted","Data":"3ee03016a1f10ac83c0527eb53986f9a904f35837aecd2fa4c6753083235f18d"} Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.795219 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.795668 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.796015 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.796470 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96","Type":"ContainerDied","Data":"c886e817586198f4cdc789e363ad7d5558146618ebce84ef0fc439c9b47902c9"} Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.796498 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c886e817586198f4cdc789e363ad7d5558146618ebce84ef0fc439c9b47902c9" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.796485 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.799153 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.799921 4904 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634" exitCode=0 Feb 14 11:14:10 crc kubenswrapper[4904]: E0214 11:14:10.800489 4904 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.17:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.810814 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.811424 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:10 crc kubenswrapper[4904]: I0214 11:14:10.811802 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:12 crc kubenswrapper[4904]: E0214 11:14:12.384907 4904 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="3.2s" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.190856 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h2tbz" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.191782 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h2tbz" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.235969 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h2tbz" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.236448 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.236909 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.238339 4904 status_manager.go:851] "Failed to get status for pod" podUID="9a1b8061-2687-4083-8e86-6c9b99daf8a5" pod="openshift-marketplace/redhat-marketplace-h2tbz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-h2tbz\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.238585 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.545905 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.546862 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.547366 4904 status_manager.go:851] "Failed to get status for pod" podUID="9a1b8061-2687-4083-8e86-6c9b99daf8a5" pod="openshift-marketplace/redhat-marketplace-h2tbz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-h2tbz\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.547696 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.548047 4904 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.548297 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.548500 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.719267 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.719354 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.719390 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.719444 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.719518 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.719616 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.719652 4904 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.719667 4904 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.816922 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.817532 4904 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37" exitCode=0 Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.818281 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.818374 4904 scope.go:117] "RemoveContainer" containerID="8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.820363 4904 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.873498 4904 scope.go:117] "RemoveContainer" containerID="0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.881644 4904 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.882681 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.883996 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.884503 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.884748 4904 status_manager.go:851] "Failed to get status for pod" podUID="9a1b8061-2687-4083-8e86-6c9b99daf8a5" pod="openshift-marketplace/redhat-marketplace-h2tbz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-h2tbz\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.884950 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.894499 4904 scope.go:117] "RemoveContainer" containerID="6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.904072 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h2tbz" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.904628 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.905177 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.905624 4904 status_manager.go:851] "Failed to get status for pod" podUID="9a1b8061-2687-4083-8e86-6c9b99daf8a5" pod="openshift-marketplace/redhat-marketplace-h2tbz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-h2tbz\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.905795 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.916193 4904 scope.go:117] "RemoveContainer" containerID="bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.948902 4904 scope.go:117] "RemoveContainer" containerID="de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.961679 4904 scope.go:117] "RemoveContainer" containerID="d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.988968 4904 scope.go:117] "RemoveContainer" containerID="8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f" Feb 14 11:14:13 crc kubenswrapper[4904]: E0214 11:14:13.989439 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\": container with ID starting with 8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f not found: ID does not exist" containerID="8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.989482 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f"} err="failed to get container status \"8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\": rpc error: code = NotFound desc = could not find container \"8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f\": container with ID starting with 8f6420969b69906e38d06effe34a42242d22e706979937c0397bd5d51498bc7f not found: ID does not exist" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.989524 4904 scope.go:117] "RemoveContainer" containerID="0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90" Feb 14 11:14:13 crc kubenswrapper[4904]: E0214 11:14:13.990039 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\": container with ID starting with 0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90 not found: ID does not exist" containerID="0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.990076 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90"} err="failed to get container status \"0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\": rpc error: code = NotFound desc = could not find container \"0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90\": container with ID starting with 0452109bde2e05a541cf9480a1b46f4323428ecdc6fff267be3d896c8aebba90 not found: ID does not exist" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.990103 4904 scope.go:117] "RemoveContainer" containerID="6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37" Feb 14 11:14:13 crc kubenswrapper[4904]: E0214 11:14:13.990528 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\": container with ID starting with 6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37 not found: ID does not exist" containerID="6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.990546 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37"} err="failed to get container status \"6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\": rpc error: code = NotFound desc = could not find container \"6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37\": container with ID starting with 6f235d63f3299c09ac6a66d835bb24e27c13e0b05dbbb279a2a58a360cf48b37 not found: ID does not exist" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.990558 4904 scope.go:117] "RemoveContainer" containerID="bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e" Feb 14 11:14:13 crc kubenswrapper[4904]: E0214 11:14:13.991156 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\": container with ID starting with bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e not found: ID does not exist" containerID="bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.991175 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e"} err="failed to get container status \"bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\": rpc error: code = NotFound desc = could not find container \"bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e\": container with ID starting with bd8934a19c9a531269d9e38f10ad8c41c3ccddc9337fd53e7df0b56d1993cb2e not found: ID does not exist" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.991187 4904 scope.go:117] "RemoveContainer" containerID="de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634" Feb 14 11:14:13 crc kubenswrapper[4904]: E0214 11:14:13.991460 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\": container with ID starting with de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634 not found: ID does not exist" containerID="de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.991484 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634"} err="failed to get container status \"de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\": rpc error: code = NotFound desc = could not find container \"de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634\": container with ID starting with de90f4d5befad27302abb3b63727e50ef47297b10e1de95ed161021654f42634 not found: ID does not exist" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.991499 4904 scope.go:117] "RemoveContainer" containerID="d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0" Feb 14 11:14:13 crc kubenswrapper[4904]: E0214 11:14:13.991755 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\": container with ID starting with d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0 not found: ID does not exist" containerID="d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0" Feb 14 11:14:13 crc kubenswrapper[4904]: I0214 11:14:13.991777 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0"} err="failed to get container status \"d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\": rpc error: code = NotFound desc = could not find container \"d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0\": container with ID starting with d2d80a3d7930f6da0e8cf9f6f604faa3694a416ba300a3c42a971ebd1f2cd9a0 not found: ID does not exist" Feb 14 11:14:14 crc kubenswrapper[4904]: I0214 11:14:14.577118 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-99bfv" Feb 14 11:14:14 crc kubenswrapper[4904]: I0214 11:14:14.577416 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-99bfv" Feb 14 11:14:14 crc kubenswrapper[4904]: I0214 11:14:14.630459 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-99bfv" Feb 14 11:14:14 crc kubenswrapper[4904]: I0214 11:14:14.631004 4904 status_manager.go:851] "Failed to get status for pod" podUID="9a1b8061-2687-4083-8e86-6c9b99daf8a5" pod="openshift-marketplace/redhat-marketplace-h2tbz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-h2tbz\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:14 crc kubenswrapper[4904]: I0214 11:14:14.631311 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:14 crc kubenswrapper[4904]: I0214 11:14:14.631666 4904 status_manager.go:851] "Failed to get status for pod" podUID="4233cb32-d700-4c17-a97b-82d5bcd6ea5f" pod="openshift-marketplace/community-operators-99bfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-99bfv\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:14 crc kubenswrapper[4904]: I0214 11:14:14.632026 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:14 crc kubenswrapper[4904]: I0214 11:14:14.632415 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:14 crc kubenswrapper[4904]: I0214 11:14:14.872247 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-99bfv" Feb 14 11:14:14 crc kubenswrapper[4904]: I0214 11:14:14.872704 4904 status_manager.go:851] "Failed to get status for pod" podUID="9a1b8061-2687-4083-8e86-6c9b99daf8a5" pod="openshift-marketplace/redhat-marketplace-h2tbz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-h2tbz\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:14 crc kubenswrapper[4904]: I0214 11:14:14.873012 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:14 crc kubenswrapper[4904]: I0214 11:14:14.873199 4904 status_manager.go:851] "Failed to get status for pod" podUID="4233cb32-d700-4c17-a97b-82d5bcd6ea5f" pod="openshift-marketplace/community-operators-99bfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-99bfv\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:14 crc kubenswrapper[4904]: I0214 11:14:14.873344 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:14 crc kubenswrapper[4904]: I0214 11:14:14.873481 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:15 crc kubenswrapper[4904]: I0214 11:14:15.584896 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4m6zf" Feb 14 11:14:15 crc kubenswrapper[4904]: I0214 11:14:15.585237 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4m6zf" Feb 14 11:14:15 crc kubenswrapper[4904]: E0214 11:14:15.585821 4904 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="6.4s" Feb 14 11:14:15 crc kubenswrapper[4904]: I0214 11:14:15.623854 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4m6zf" Feb 14 11:14:15 crc kubenswrapper[4904]: I0214 11:14:15.624286 4904 status_manager.go:851] "Failed to get status for pod" podUID="4233cb32-d700-4c17-a97b-82d5bcd6ea5f" pod="openshift-marketplace/community-operators-99bfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-99bfv\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:15 crc kubenswrapper[4904]: I0214 11:14:15.624465 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:15 crc kubenswrapper[4904]: I0214 11:14:15.624777 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:15 crc kubenswrapper[4904]: I0214 11:14:15.624940 4904 status_manager.go:851] "Failed to get status for pod" podUID="9a1b8061-2687-4083-8e86-6c9b99daf8a5" pod="openshift-marketplace/redhat-marketplace-h2tbz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-h2tbz\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:15 crc kubenswrapper[4904]: I0214 11:14:15.625094 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:15 crc kubenswrapper[4904]: I0214 11:14:15.876214 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4m6zf" Feb 14 11:14:15 crc kubenswrapper[4904]: I0214 11:14:15.876628 4904 status_manager.go:851] "Failed to get status for pod" podUID="4233cb32-d700-4c17-a97b-82d5bcd6ea5f" pod="openshift-marketplace/community-operators-99bfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-99bfv\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:15 crc kubenswrapper[4904]: I0214 11:14:15.877026 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:15 crc kubenswrapper[4904]: I0214 11:14:15.877204 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:15 crc kubenswrapper[4904]: I0214 11:14:15.877347 4904 status_manager.go:851] "Failed to get status for pod" podUID="9a1b8061-2687-4083-8e86-6c9b99daf8a5" pod="openshift-marketplace/redhat-marketplace-h2tbz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-h2tbz\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:15 crc kubenswrapper[4904]: I0214 11:14:15.877491 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:16 crc kubenswrapper[4904]: I0214 11:14:16.974534 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wxxxn" Feb 14 11:14:16 crc kubenswrapper[4904]: I0214 11:14:16.974581 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wxxxn" Feb 14 11:14:17 crc kubenswrapper[4904]: I0214 11:14:17.020525 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wxxxn" Feb 14 11:14:17 crc kubenswrapper[4904]: I0214 11:14:17.021034 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:17 crc kubenswrapper[4904]: I0214 11:14:17.021487 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:17 crc kubenswrapper[4904]: I0214 11:14:17.022217 4904 status_manager.go:851] "Failed to get status for pod" podUID="9a1b8061-2687-4083-8e86-6c9b99daf8a5" pod="openshift-marketplace/redhat-marketplace-h2tbz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-h2tbz\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:17 crc kubenswrapper[4904]: I0214 11:14:17.022698 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:17 crc kubenswrapper[4904]: I0214 11:14:17.023163 4904 status_manager.go:851] "Failed to get status for pod" podUID="4233cb32-d700-4c17-a97b-82d5bcd6ea5f" pod="openshift-marketplace/community-operators-99bfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-99bfv\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:17 crc kubenswrapper[4904]: I0214 11:14:17.890302 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wxxxn" Feb 14 11:14:17 crc kubenswrapper[4904]: I0214 11:14:17.890960 4904 status_manager.go:851] "Failed to get status for pod" podUID="4233cb32-d700-4c17-a97b-82d5bcd6ea5f" pod="openshift-marketplace/community-operators-99bfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-99bfv\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:17 crc kubenswrapper[4904]: I0214 11:14:17.891327 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:17 crc kubenswrapper[4904]: I0214 11:14:17.891650 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:17 crc kubenswrapper[4904]: I0214 11:14:17.891955 4904 status_manager.go:851] "Failed to get status for pod" podUID="9a1b8061-2687-4083-8e86-6c9b99daf8a5" pod="openshift-marketplace/redhat-marketplace-h2tbz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-h2tbz\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:17 crc kubenswrapper[4904]: I0214 11:14:17.892242 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:19 crc kubenswrapper[4904]: I0214 11:14:19.837983 4904 status_manager.go:851] "Failed to get status for pod" podUID="9a1b8061-2687-4083-8e86-6c9b99daf8a5" pod="openshift-marketplace/redhat-marketplace-h2tbz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-h2tbz\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:19 crc kubenswrapper[4904]: I0214 11:14:19.838749 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:19 crc kubenswrapper[4904]: I0214 11:14:19.839044 4904 status_manager.go:851] "Failed to get status for pod" podUID="4233cb32-d700-4c17-a97b-82d5bcd6ea5f" pod="openshift-marketplace/community-operators-99bfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-99bfv\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:19 crc kubenswrapper[4904]: I0214 11:14:19.839308 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:19 crc kubenswrapper[4904]: I0214 11:14:19.839596 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:19 crc kubenswrapper[4904]: E0214 11:14:19.869194 4904 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.17:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-operators-wxxxn.189418a4804152bc openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-operators-wxxxn,UID:6b349538-4188-447c-aee9-74797e324398,APIVersion:v1,ResourceVersion:29663,FieldPath:spec.initContainers{extract-content},},Reason:Pulled,Message:Successfully pulled image \"registry.redhat.io/redhat/redhat-operator-index:v4.18\" in 551ms (551ms including waiting). Image size: 1701129928 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-14 11:14:08.268915388 +0000 UTC m=+239.081680049,LastTimestamp:2026-02-14 11:14:08.268915388 +0000 UTC m=+239.081680049,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 14 11:14:20 crc kubenswrapper[4904]: E0214 11:14:20.252787 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:14:20Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:14:20Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:14:20Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-14T11:14:20Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:20 crc kubenswrapper[4904]: E0214 11:14:20.253082 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:20 crc kubenswrapper[4904]: E0214 11:14:20.253404 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:20 crc kubenswrapper[4904]: E0214 11:14:20.254424 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:20 crc kubenswrapper[4904]: E0214 11:14:20.254627 4904 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:20 crc kubenswrapper[4904]: E0214 11:14:20.254652 4904 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 14 11:14:20 crc kubenswrapper[4904]: I0214 11:14:20.836972 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:20 crc kubenswrapper[4904]: I0214 11:14:20.837690 4904 status_manager.go:851] "Failed to get status for pod" podUID="9a1b8061-2687-4083-8e86-6c9b99daf8a5" pod="openshift-marketplace/redhat-marketplace-h2tbz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-h2tbz\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:20 crc kubenswrapper[4904]: I0214 11:14:20.838207 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:20 crc kubenswrapper[4904]: I0214 11:14:20.838645 4904 status_manager.go:851] "Failed to get status for pod" podUID="4233cb32-d700-4c17-a97b-82d5bcd6ea5f" pod="openshift-marketplace/community-operators-99bfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-99bfv\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:20 crc kubenswrapper[4904]: I0214 11:14:20.839034 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:20 crc kubenswrapper[4904]: I0214 11:14:20.839285 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:20 crc kubenswrapper[4904]: I0214 11:14:20.850526 4904 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ed38d346-9528-40d4-96fd-ad2476654244" Feb 14 11:14:20 crc kubenswrapper[4904]: I0214 11:14:20.850555 4904 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ed38d346-9528-40d4-96fd-ad2476654244" Feb 14 11:14:20 crc kubenswrapper[4904]: E0214 11:14:20.850889 4904 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:20 crc kubenswrapper[4904]: I0214 11:14:20.851458 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:20 crc kubenswrapper[4904]: W0214 11:14:20.877988 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-ce8de3c40dceebd6c4e0f2e280cd3f29acca678c303789a9ce6c6db8cd2edec6 WatchSource:0}: Error finding container ce8de3c40dceebd6c4e0f2e280cd3f29acca678c303789a9ce6c6db8cd2edec6: Status 404 returned error can't find the container with id ce8de3c40dceebd6c4e0f2e280cd3f29acca678c303789a9ce6c6db8cd2edec6 Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.872433 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.872682 4904 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2" exitCode=1 Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.872710 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2"} Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.873089 4904 scope.go:117] "RemoveContainer" containerID="096c343d334cc70b8519dbb4a02820b4554673b7ac83e06916085f4347ed61b2" Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.873427 4904 status_manager.go:851] "Failed to get status for pod" podUID="9a1b8061-2687-4083-8e86-6c9b99daf8a5" pod="openshift-marketplace/redhat-marketplace-h2tbz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-h2tbz\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.873670 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.873917 4904 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.873972 4904 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="669409455efd517ee4c55a862a0daed5fc4bad186d65e51c7793d460e86ab3aa" exitCode=0 Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.873996 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"669409455efd517ee4c55a862a0daed5fc4bad186d65e51c7793d460e86ab3aa"} Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.874013 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ce8de3c40dceebd6c4e0f2e280cd3f29acca678c303789a9ce6c6db8cd2edec6"} Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.874144 4904 status_manager.go:851] "Failed to get status for pod" podUID="4233cb32-d700-4c17-a97b-82d5bcd6ea5f" pod="openshift-marketplace/community-operators-99bfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-99bfv\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.874210 4904 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ed38d346-9528-40d4-96fd-ad2476654244" Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.874225 4904 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ed38d346-9528-40d4-96fd-ad2476654244" Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.874371 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:21 crc kubenswrapper[4904]: E0214 11:14:21.874457 4904 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.874540 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.874767 4904 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.875064 4904 status_manager.go:851] "Failed to get status for pod" podUID="4233cb32-d700-4c17-a97b-82d5bcd6ea5f" pod="openshift-marketplace/community-operators-99bfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-99bfv\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.875289 4904 status_manager.go:851] "Failed to get status for pod" podUID="6b349538-4188-447c-aee9-74797e324398" pod="openshift-marketplace/redhat-operators-wxxxn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-wxxxn\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.875503 4904 status_manager.go:851] "Failed to get status for pod" podUID="dfb9b35f-85bb-43ec-a29d-897b3ebf0da9" pod="openshift-marketplace/certified-operators-4m6zf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-4m6zf\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.875824 4904 status_manager.go:851] "Failed to get status for pod" podUID="9a1b8061-2687-4083-8e86-6c9b99daf8a5" pod="openshift-marketplace/redhat-marketplace-h2tbz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-h2tbz\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:21 crc kubenswrapper[4904]: I0214 11:14:21.876153 4904 status_manager.go:851] "Failed to get status for pod" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Feb 14 11:14:21 crc kubenswrapper[4904]: E0214 11:14:21.986981 4904 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="7s" Feb 14 11:14:22 crc kubenswrapper[4904]: I0214 11:14:22.886195 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 14 11:14:22 crc kubenswrapper[4904]: I0214 11:14:22.886556 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6c3bb032151581684f029b444cb60755bad5a3627048db4ba5329a27251b5152"} Feb 14 11:14:22 crc kubenswrapper[4904]: I0214 11:14:22.889200 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3c88b7515f2f682695d397cc8923ede4f2c1fad9206979228bc545be82198f7a"} Feb 14 11:14:22 crc kubenswrapper[4904]: I0214 11:14:22.889245 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"07debdbf1605460f27371abefd166f136fefb769d1a15ef479a377efd07a178b"} Feb 14 11:14:22 crc kubenswrapper[4904]: I0214 11:14:22.889258 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5b60b2d8eb848669e7b262174323ec3ea7ee76d031d13d9aa64270ee7b6db1f3"} Feb 14 11:14:22 crc kubenswrapper[4904]: I0214 11:14:22.889269 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ded6ed365f2be2c7597247acd0b837df1aec7b3bf77dd4cb9cecf2deee250100"} Feb 14 11:14:23 crc kubenswrapper[4904]: I0214 11:14:23.896992 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7740961e0e9f252b80e5ef08bbe46ea30822bec957bfd54e608cc31a754a25b5"} Feb 14 11:14:23 crc kubenswrapper[4904]: I0214 11:14:23.897366 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:23 crc kubenswrapper[4904]: I0214 11:14:23.897242 4904 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ed38d346-9528-40d4-96fd-ad2476654244" Feb 14 11:14:23 crc kubenswrapper[4904]: I0214 11:14:23.897389 4904 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ed38d346-9528-40d4-96fd-ad2476654244" Feb 14 11:14:25 crc kubenswrapper[4904]: I0214 11:14:25.287965 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:14:25 crc kubenswrapper[4904]: I0214 11:14:25.851792 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:25 crc kubenswrapper[4904]: I0214 11:14:25.852111 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:25 crc kubenswrapper[4904]: I0214 11:14:25.857274 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:28 crc kubenswrapper[4904]: I0214 11:14:28.905701 4904 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:28 crc kubenswrapper[4904]: I0214 11:14:28.920485 4904 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ed38d346-9528-40d4-96fd-ad2476654244" Feb 14 11:14:28 crc kubenswrapper[4904]: I0214 11:14:28.920512 4904 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ed38d346-9528-40d4-96fd-ad2476654244" Feb 14 11:14:28 crc kubenswrapper[4904]: I0214 11:14:28.932662 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:29 crc kubenswrapper[4904]: I0214 11:14:29.089748 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:14:29 crc kubenswrapper[4904]: I0214 11:14:29.093275 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:14:29 crc kubenswrapper[4904]: I0214 11:14:29.856933 4904 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="1f82b130-2f8d-4e6a-8b10-5e96c7326f53" Feb 14 11:14:29 crc kubenswrapper[4904]: I0214 11:14:29.926269 4904 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ed38d346-9528-40d4-96fd-ad2476654244" Feb 14 11:14:29 crc kubenswrapper[4904]: I0214 11:14:29.926302 4904 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ed38d346-9528-40d4-96fd-ad2476654244" Feb 14 11:14:29 crc kubenswrapper[4904]: I0214 11:14:29.930776 4904 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="1f82b130-2f8d-4e6a-8b10-5e96c7326f53" Feb 14 11:14:35 crc kubenswrapper[4904]: I0214 11:14:35.291868 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 14 11:14:38 crc kubenswrapper[4904]: I0214 11:14:38.806332 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 14 11:14:39 crc kubenswrapper[4904]: I0214 11:14:39.275165 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 14 11:14:39 crc kubenswrapper[4904]: I0214 11:14:39.317452 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 14 11:14:39 crc kubenswrapper[4904]: I0214 11:14:39.438154 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 14 11:14:39 crc kubenswrapper[4904]: I0214 11:14:39.467053 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 14 11:14:39 crc kubenswrapper[4904]: I0214 11:14:39.730581 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 14 11:14:39 crc kubenswrapper[4904]: I0214 11:14:39.972921 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 14 11:14:40 crc kubenswrapper[4904]: I0214 11:14:40.179028 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 14 11:14:40 crc kubenswrapper[4904]: I0214 11:14:40.437460 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 14 11:14:40 crc kubenswrapper[4904]: I0214 11:14:40.578423 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 14 11:14:40 crc kubenswrapper[4904]: I0214 11:14:40.670705 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 14 11:14:40 crc kubenswrapper[4904]: I0214 11:14:40.739381 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 14 11:14:40 crc kubenswrapper[4904]: I0214 11:14:40.826366 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 14 11:14:40 crc kubenswrapper[4904]: I0214 11:14:40.884278 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 14 11:14:40 crc kubenswrapper[4904]: I0214 11:14:40.921033 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 14 11:14:40 crc kubenswrapper[4904]: I0214 11:14:40.972527 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 14 11:14:41 crc kubenswrapper[4904]: I0214 11:14:41.179908 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 14 11:14:41 crc kubenswrapper[4904]: I0214 11:14:41.330337 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 14 11:14:41 crc kubenswrapper[4904]: I0214 11:14:41.460615 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 14 11:14:41 crc kubenswrapper[4904]: I0214 11:14:41.461722 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 14 11:14:41 crc kubenswrapper[4904]: I0214 11:14:41.559933 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 14 11:14:41 crc kubenswrapper[4904]: I0214 11:14:41.624001 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 14 11:14:41 crc kubenswrapper[4904]: I0214 11:14:41.721368 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 14 11:14:41 crc kubenswrapper[4904]: I0214 11:14:41.823715 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 14 11:14:41 crc kubenswrapper[4904]: I0214 11:14:41.859761 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 14 11:14:41 crc kubenswrapper[4904]: I0214 11:14:41.895128 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 14 11:14:41 crc kubenswrapper[4904]: I0214 11:14:41.942569 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 14 11:14:41 crc kubenswrapper[4904]: I0214 11:14:41.956425 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 14 11:14:42 crc kubenswrapper[4904]: I0214 11:14:42.174773 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 14 11:14:42 crc kubenswrapper[4904]: I0214 11:14:42.357170 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 14 11:14:42 crc kubenswrapper[4904]: I0214 11:14:42.363253 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 14 11:14:42 crc kubenswrapper[4904]: I0214 11:14:42.420249 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 14 11:14:42 crc kubenswrapper[4904]: I0214 11:14:42.454728 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 14 11:14:42 crc kubenswrapper[4904]: I0214 11:14:42.774650 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 14 11:14:42 crc kubenswrapper[4904]: I0214 11:14:42.799701 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 14 11:14:42 crc kubenswrapper[4904]: I0214 11:14:42.829299 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 14 11:14:42 crc kubenswrapper[4904]: I0214 11:14:42.859514 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 14 11:14:42 crc kubenswrapper[4904]: I0214 11:14:42.965194 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.080577 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.242119 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.245527 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.270771 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.324281 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.331307 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.418872 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.461631 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.526437 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.555487 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.598626 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.617433 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.641220 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.666133 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.688045 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.723430 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.824011 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.899096 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.908655 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.908925 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 14 11:14:43 crc kubenswrapper[4904]: I0214 11:14:43.989242 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.035948 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.058163 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.278169 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.320968 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.344438 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.351520 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.356288 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.452472 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.491365 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.513788 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.616502 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.640641 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.708078 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.804744 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.809693 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.841146 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.842210 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.843345 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.863960 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.902166 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.906605 4904 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 14 11:14:44 crc kubenswrapper[4904]: I0214 11:14:44.957742 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.014777 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.063513 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.064162 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.075692 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.100133 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.174052 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.208659 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.219672 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.329915 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.338786 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.408923 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.422924 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.487625 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.533134 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.554920 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.560428 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.621647 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.697006 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.736722 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.758757 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.773145 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.780584 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.783625 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.834515 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 14 11:14:45 crc kubenswrapper[4904]: I0214 11:14:45.966159 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.027468 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.028038 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.036092 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.052992 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.077275 4904 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.162382 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.163323 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.197421 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.207115 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.289685 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.323201 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.336913 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.344865 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.436026 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.438537 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.447184 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.458375 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.566181 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.710341 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.726423 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.754742 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.772923 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.896432 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.928121 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 14 11:14:46 crc kubenswrapper[4904]: I0214 11:14:46.969620 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 14 11:14:47 crc kubenswrapper[4904]: I0214 11:14:47.092398 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 14 11:14:47 crc kubenswrapper[4904]: I0214 11:14:47.215584 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 14 11:14:47 crc kubenswrapper[4904]: I0214 11:14:47.255905 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 14 11:14:47 crc kubenswrapper[4904]: I0214 11:14:47.303578 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 14 11:14:47 crc kubenswrapper[4904]: I0214 11:14:47.363156 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 14 11:14:47 crc kubenswrapper[4904]: I0214 11:14:47.381332 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 14 11:14:47 crc kubenswrapper[4904]: I0214 11:14:47.571784 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 14 11:14:47 crc kubenswrapper[4904]: I0214 11:14:47.593716 4904 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 14 11:14:47 crc kubenswrapper[4904]: I0214 11:14:47.694157 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 14 11:14:47 crc kubenswrapper[4904]: I0214 11:14:47.713757 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 14 11:14:47 crc kubenswrapper[4904]: I0214 11:14:47.736698 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 14 11:14:47 crc kubenswrapper[4904]: I0214 11:14:47.751327 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 14 11:14:48 crc kubenswrapper[4904]: I0214 11:14:48.102069 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 14 11:14:48 crc kubenswrapper[4904]: I0214 11:14:48.196486 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 14 11:14:48 crc kubenswrapper[4904]: I0214 11:14:48.234222 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 14 11:14:48 crc kubenswrapper[4904]: I0214 11:14:48.348590 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 14 11:14:48 crc kubenswrapper[4904]: I0214 11:14:48.453385 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 14 11:14:48 crc kubenswrapper[4904]: I0214 11:14:48.923850 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 14 11:14:48 crc kubenswrapper[4904]: I0214 11:14:48.941816 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 14 11:14:48 crc kubenswrapper[4904]: I0214 11:14:48.948625 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 14 11:14:48 crc kubenswrapper[4904]: I0214 11:14:48.967962 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.077723 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.169211 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.176685 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.242427 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.287160 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.295031 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.335972 4904 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.338851 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4m6zf" podStartSLOduration=42.972593634 podStartE2EDuration="44.338813616s" podCreationTimestamp="2026-02-14 11:14:05 +0000 UTC" firstStartedPulling="2026-02-14 11:14:06.704481457 +0000 UTC m=+237.517246118" lastFinishedPulling="2026-02-14 11:14:08.070701439 +0000 UTC m=+238.883466100" observedRunningTime="2026-02-14 11:14:28.743886244 +0000 UTC m=+259.556650905" watchObservedRunningTime="2026-02-14 11:14:49.338813616 +0000 UTC m=+280.151578287" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.340385 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wxxxn" podStartSLOduration=40.920321807 podStartE2EDuration="43.340378711s" podCreationTimestamp="2026-02-14 11:14:06 +0000 UTC" firstStartedPulling="2026-02-14 11:14:07.71701855 +0000 UTC m=+238.529783211" lastFinishedPulling="2026-02-14 11:14:10.137075454 +0000 UTC m=+240.949840115" observedRunningTime="2026-02-14 11:14:28.728562634 +0000 UTC m=+259.541327285" watchObservedRunningTime="2026-02-14 11:14:49.340378711 +0000 UTC m=+280.153143382" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.341216 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.341270 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.341637 4904 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ed38d346-9528-40d4-96fd-ad2476654244" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.341667 4904 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ed38d346-9528-40d4-96fd-ad2476654244" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.347019 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.362417 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=21.362402559 podStartE2EDuration="21.362402559s" podCreationTimestamp="2026-02-14 11:14:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:14:49.360549276 +0000 UTC m=+280.173313957" watchObservedRunningTime="2026-02-14 11:14:49.362402559 +0000 UTC m=+280.175167210" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.475156 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.587102 4904 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.681937 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.691725 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.724626 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.781898 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.811129 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.934055 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.953526 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.987478 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 14 11:14:49 crc kubenswrapper[4904]: I0214 11:14:49.992148 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 14 11:14:50 crc kubenswrapper[4904]: I0214 11:14:50.055119 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 14 11:14:50 crc kubenswrapper[4904]: I0214 11:14:50.114504 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 14 11:14:50 crc kubenswrapper[4904]: I0214 11:14:50.140502 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 14 11:14:50 crc kubenswrapper[4904]: I0214 11:14:50.296638 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 14 11:14:50 crc kubenswrapper[4904]: I0214 11:14:50.329620 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 14 11:14:50 crc kubenswrapper[4904]: I0214 11:14:50.359224 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 14 11:14:50 crc kubenswrapper[4904]: I0214 11:14:50.436942 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 14 11:14:50 crc kubenswrapper[4904]: I0214 11:14:50.443950 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 14 11:14:50 crc kubenswrapper[4904]: I0214 11:14:50.460052 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 14 11:14:50 crc kubenswrapper[4904]: I0214 11:14:50.725112 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 14 11:14:50 crc kubenswrapper[4904]: I0214 11:14:50.744511 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 14 11:14:50 crc kubenswrapper[4904]: I0214 11:14:50.781092 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 14 11:14:50 crc kubenswrapper[4904]: I0214 11:14:50.793347 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 14 11:14:50 crc kubenswrapper[4904]: I0214 11:14:50.861504 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 14 11:14:50 crc kubenswrapper[4904]: I0214 11:14:50.933943 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 14 11:14:51 crc kubenswrapper[4904]: I0214 11:14:51.165785 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 14 11:14:51 crc kubenswrapper[4904]: I0214 11:14:51.166657 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 14 11:14:51 crc kubenswrapper[4904]: I0214 11:14:51.175702 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 14 11:14:51 crc kubenswrapper[4904]: I0214 11:14:51.313573 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 14 11:14:51 crc kubenswrapper[4904]: I0214 11:14:51.329674 4904 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 14 11:14:51 crc kubenswrapper[4904]: I0214 11:14:51.329942 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://11c735ef5b3fc38cf3d8214a4c0626433da291ef89e18eec6404498f2bea2c4c" gracePeriod=5 Feb 14 11:14:51 crc kubenswrapper[4904]: I0214 11:14:51.398057 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 14 11:14:51 crc kubenswrapper[4904]: I0214 11:14:51.615853 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 14 11:14:51 crc kubenswrapper[4904]: I0214 11:14:51.616201 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 14 11:14:51 crc kubenswrapper[4904]: I0214 11:14:51.619023 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 14 11:14:51 crc kubenswrapper[4904]: I0214 11:14:51.664136 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 14 11:14:51 crc kubenswrapper[4904]: I0214 11:14:51.723649 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 14 11:14:51 crc kubenswrapper[4904]: I0214 11:14:51.826094 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 14 11:14:51 crc kubenswrapper[4904]: I0214 11:14:51.859233 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 14 11:14:51 crc kubenswrapper[4904]: I0214 11:14:51.970580 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 14 11:14:51 crc kubenswrapper[4904]: I0214 11:14:51.970720 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.024347 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.024659 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.030530 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.039862 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.040211 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.133548 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.135679 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.158220 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.277316 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.329518 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.372169 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.450791 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.456417 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.458821 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.479189 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.495370 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.587068 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.593174 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.618923 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.657072 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.749787 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.782218 4904 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.783629 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.869515 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 14 11:14:52 crc kubenswrapper[4904]: I0214 11:14:52.950015 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 14 11:14:53 crc kubenswrapper[4904]: I0214 11:14:53.075538 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 14 11:14:53 crc kubenswrapper[4904]: I0214 11:14:53.083380 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 14 11:14:53 crc kubenswrapper[4904]: I0214 11:14:53.088821 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 14 11:14:53 crc kubenswrapper[4904]: I0214 11:14:53.182594 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 14 11:14:53 crc kubenswrapper[4904]: I0214 11:14:53.331168 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 14 11:14:53 crc kubenswrapper[4904]: I0214 11:14:53.458140 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 14 11:14:53 crc kubenswrapper[4904]: I0214 11:14:53.574332 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 14 11:14:53 crc kubenswrapper[4904]: I0214 11:14:53.577892 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 14 11:14:53 crc kubenswrapper[4904]: I0214 11:14:53.825585 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 14 11:14:53 crc kubenswrapper[4904]: I0214 11:14:53.842222 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 14 11:14:54 crc kubenswrapper[4904]: I0214 11:14:54.010730 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 14 11:14:54 crc kubenswrapper[4904]: I0214 11:14:54.170796 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 14 11:14:54 crc kubenswrapper[4904]: I0214 11:14:54.451946 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 14 11:14:54 crc kubenswrapper[4904]: I0214 11:14:54.566030 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 14 11:14:54 crc kubenswrapper[4904]: I0214 11:14:54.631005 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 14 11:14:54 crc kubenswrapper[4904]: I0214 11:14:54.742326 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 14 11:14:55 crc kubenswrapper[4904]: I0214 11:14:55.012537 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 14 11:14:55 crc kubenswrapper[4904]: I0214 11:14:55.015640 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 14 11:14:55 crc kubenswrapper[4904]: I0214 11:14:55.078070 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 14 11:14:55 crc kubenswrapper[4904]: I0214 11:14:55.133020 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 14 11:14:55 crc kubenswrapper[4904]: I0214 11:14:55.421129 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 14 11:14:55 crc kubenswrapper[4904]: I0214 11:14:55.543539 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 14 11:14:55 crc kubenswrapper[4904]: I0214 11:14:55.765462 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 14 11:14:55 crc kubenswrapper[4904]: I0214 11:14:55.981142 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 14 11:14:56 crc kubenswrapper[4904]: I0214 11:14:56.033989 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 14 11:14:56 crc kubenswrapper[4904]: I0214 11:14:56.142742 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 14 11:14:56 crc kubenswrapper[4904]: I0214 11:14:56.449612 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 14 11:14:56 crc kubenswrapper[4904]: I0214 11:14:56.915767 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 14 11:14:56 crc kubenswrapper[4904]: I0214 11:14:56.915861 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:56 crc kubenswrapper[4904]: I0214 11:14:56.939302 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.060312 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.060380 4904 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="11c735ef5b3fc38cf3d8214a4c0626433da291ef89e18eec6404498f2bea2c4c" exitCode=137 Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.060448 4904 scope.go:117] "RemoveContainer" containerID="11c735ef5b3fc38cf3d8214a4c0626433da291ef89e18eec6404498f2bea2c4c" Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.060604 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.076264 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.076325 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.076443 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.076499 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.076528 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.076718 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.076796 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.076752 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.076773 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.078230 4904 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.078249 4904 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.078260 4904 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.078358 4904 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.083332 4904 scope.go:117] "RemoveContainer" containerID="11c735ef5b3fc38cf3d8214a4c0626433da291ef89e18eec6404498f2bea2c4c" Feb 14 11:14:57 crc kubenswrapper[4904]: E0214 11:14:57.083800 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11c735ef5b3fc38cf3d8214a4c0626433da291ef89e18eec6404498f2bea2c4c\": container with ID starting with 11c735ef5b3fc38cf3d8214a4c0626433da291ef89e18eec6404498f2bea2c4c not found: ID does not exist" containerID="11c735ef5b3fc38cf3d8214a4c0626433da291ef89e18eec6404498f2bea2c4c" Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.083850 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11c735ef5b3fc38cf3d8214a4c0626433da291ef89e18eec6404498f2bea2c4c"} err="failed to get container status \"11c735ef5b3fc38cf3d8214a4c0626433da291ef89e18eec6404498f2bea2c4c\": rpc error: code = NotFound desc = could not find container \"11c735ef5b3fc38cf3d8214a4c0626433da291ef89e18eec6404498f2bea2c4c\": container with ID starting with 11c735ef5b3fc38cf3d8214a4c0626433da291ef89e18eec6404498f2bea2c4c not found: ID does not exist" Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.092618 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.179611 4904 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 14 11:14:57 crc kubenswrapper[4904]: I0214 11:14:57.842524 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.164060 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7"] Feb 14 11:15:00 crc kubenswrapper[4904]: E0214 11:15:00.164570 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.164585 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 14 11:15:00 crc kubenswrapper[4904]: E0214 11:15:00.164602 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" containerName="installer" Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.164609 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" containerName="installer" Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.164714 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.164735 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7f3f2e1-f0e9-4eff-b4fe-8471636d8d96" containerName="installer" Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.165141 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7" Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.167699 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.172641 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.180290 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7"] Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.209749 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e31d6f5-239d-4418-b95d-dfe790798140-config-volume\") pod \"collect-profiles-29517795-5qrh7\" (UID: \"4e31d6f5-239d-4418-b95d-dfe790798140\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7" Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.209794 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stpj6\" (UniqueName: \"kubernetes.io/projected/4e31d6f5-239d-4418-b95d-dfe790798140-kube-api-access-stpj6\") pod \"collect-profiles-29517795-5qrh7\" (UID: \"4e31d6f5-239d-4418-b95d-dfe790798140\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7" Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.209825 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e31d6f5-239d-4418-b95d-dfe790798140-secret-volume\") pod \"collect-profiles-29517795-5qrh7\" (UID: \"4e31d6f5-239d-4418-b95d-dfe790798140\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7" Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.310588 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e31d6f5-239d-4418-b95d-dfe790798140-config-volume\") pod \"collect-profiles-29517795-5qrh7\" (UID: \"4e31d6f5-239d-4418-b95d-dfe790798140\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7" Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.310639 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stpj6\" (UniqueName: \"kubernetes.io/projected/4e31d6f5-239d-4418-b95d-dfe790798140-kube-api-access-stpj6\") pod \"collect-profiles-29517795-5qrh7\" (UID: \"4e31d6f5-239d-4418-b95d-dfe790798140\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7" Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.310680 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e31d6f5-239d-4418-b95d-dfe790798140-secret-volume\") pod \"collect-profiles-29517795-5qrh7\" (UID: \"4e31d6f5-239d-4418-b95d-dfe790798140\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7" Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.311582 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e31d6f5-239d-4418-b95d-dfe790798140-config-volume\") pod \"collect-profiles-29517795-5qrh7\" (UID: \"4e31d6f5-239d-4418-b95d-dfe790798140\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7" Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.320489 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e31d6f5-239d-4418-b95d-dfe790798140-secret-volume\") pod \"collect-profiles-29517795-5qrh7\" (UID: \"4e31d6f5-239d-4418-b95d-dfe790798140\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7" Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.331546 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stpj6\" (UniqueName: \"kubernetes.io/projected/4e31d6f5-239d-4418-b95d-dfe790798140-kube-api-access-stpj6\") pod \"collect-profiles-29517795-5qrh7\" (UID: \"4e31d6f5-239d-4418-b95d-dfe790798140\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7" Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.480671 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7" Feb 14 11:15:00 crc kubenswrapper[4904]: I0214 11:15:00.847097 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7"] Feb 14 11:15:01 crc kubenswrapper[4904]: I0214 11:15:01.080522 4904 generic.go:334] "Generic (PLEG): container finished" podID="4e31d6f5-239d-4418-b95d-dfe790798140" containerID="0d5480a1226127a3e228870d0f6626435aa0a80ccd0821202d2f50a4b4e22b49" exitCode=0 Feb 14 11:15:01 crc kubenswrapper[4904]: I0214 11:15:01.080580 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7" event={"ID":"4e31d6f5-239d-4418-b95d-dfe790798140","Type":"ContainerDied","Data":"0d5480a1226127a3e228870d0f6626435aa0a80ccd0821202d2f50a4b4e22b49"} Feb 14 11:15:01 crc kubenswrapper[4904]: I0214 11:15:01.080607 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7" event={"ID":"4e31d6f5-239d-4418-b95d-dfe790798140","Type":"ContainerStarted","Data":"3fd9fdb5c54e09c8fdabcd2aaba0ded3df0e268e29bf6c5d58d11e4d8b646d85"} Feb 14 11:15:02 crc kubenswrapper[4904]: I0214 11:15:02.323637 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7" Feb 14 11:15:02 crc kubenswrapper[4904]: I0214 11:15:02.434027 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e31d6f5-239d-4418-b95d-dfe790798140-secret-volume\") pod \"4e31d6f5-239d-4418-b95d-dfe790798140\" (UID: \"4e31d6f5-239d-4418-b95d-dfe790798140\") " Feb 14 11:15:02 crc kubenswrapper[4904]: I0214 11:15:02.434070 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e31d6f5-239d-4418-b95d-dfe790798140-config-volume\") pod \"4e31d6f5-239d-4418-b95d-dfe790798140\" (UID: \"4e31d6f5-239d-4418-b95d-dfe790798140\") " Feb 14 11:15:02 crc kubenswrapper[4904]: I0214 11:15:02.434160 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stpj6\" (UniqueName: \"kubernetes.io/projected/4e31d6f5-239d-4418-b95d-dfe790798140-kube-api-access-stpj6\") pod \"4e31d6f5-239d-4418-b95d-dfe790798140\" (UID: \"4e31d6f5-239d-4418-b95d-dfe790798140\") " Feb 14 11:15:02 crc kubenswrapper[4904]: I0214 11:15:02.434733 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e31d6f5-239d-4418-b95d-dfe790798140-config-volume" (OuterVolumeSpecName: "config-volume") pod "4e31d6f5-239d-4418-b95d-dfe790798140" (UID: "4e31d6f5-239d-4418-b95d-dfe790798140"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:15:02 crc kubenswrapper[4904]: I0214 11:15:02.438369 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e31d6f5-239d-4418-b95d-dfe790798140-kube-api-access-stpj6" (OuterVolumeSpecName: "kube-api-access-stpj6") pod "4e31d6f5-239d-4418-b95d-dfe790798140" (UID: "4e31d6f5-239d-4418-b95d-dfe790798140"). InnerVolumeSpecName "kube-api-access-stpj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:15:02 crc kubenswrapper[4904]: I0214 11:15:02.438665 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e31d6f5-239d-4418-b95d-dfe790798140-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4e31d6f5-239d-4418-b95d-dfe790798140" (UID: "4e31d6f5-239d-4418-b95d-dfe790798140"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:15:02 crc kubenswrapper[4904]: I0214 11:15:02.535564 4904 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e31d6f5-239d-4418-b95d-dfe790798140-config-volume\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:02 crc kubenswrapper[4904]: I0214 11:15:02.535590 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stpj6\" (UniqueName: \"kubernetes.io/projected/4e31d6f5-239d-4418-b95d-dfe790798140-kube-api-access-stpj6\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:02 crc kubenswrapper[4904]: I0214 11:15:02.535601 4904 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e31d6f5-239d-4418-b95d-dfe790798140-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:03 crc kubenswrapper[4904]: I0214 11:15:03.093370 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7" event={"ID":"4e31d6f5-239d-4418-b95d-dfe790798140","Type":"ContainerDied","Data":"3fd9fdb5c54e09c8fdabcd2aaba0ded3df0e268e29bf6c5d58d11e4d8b646d85"} Feb 14 11:15:03 crc kubenswrapper[4904]: I0214 11:15:03.093683 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fd9fdb5c54e09c8fdabcd2aaba0ded3df0e268e29bf6c5d58d11e4d8b646d85" Feb 14 11:15:03 crc kubenswrapper[4904]: I0214 11:15:03.093449 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7" Feb 14 11:15:09 crc kubenswrapper[4904]: I0214 11:15:09.676107 4904 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Feb 14 11:15:37 crc kubenswrapper[4904]: I0214 11:15:37.390003 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52"] Feb 14 11:15:37 crc kubenswrapper[4904]: I0214 11:15:37.390611 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" podUID="d910fb92-a40a-4167-be6f-113f22c206b2" containerName="route-controller-manager" containerID="cri-o://976a29dc369aa3b647f0f609fcfc3b5caccfdfc11a27fda35184577280c3470b" gracePeriod=30 Feb 14 11:15:37 crc kubenswrapper[4904]: I0214 11:15:37.480108 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5lwz5"] Feb 14 11:15:37 crc kubenswrapper[4904]: I0214 11:15:37.480300 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" podUID="5a537ec4-ab86-4bc9-81d2-3870472b767a" containerName="controller-manager" containerID="cri-o://9969e557a55d34107e04e866c201d4d8d5e8d7eb9f45fd44bde1c7338015a20e" gracePeriod=30 Feb 14 11:15:37 crc kubenswrapper[4904]: I0214 11:15:37.772905 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" Feb 14 11:15:37 crc kubenswrapper[4904]: I0214 11:15:37.846361 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:15:37 crc kubenswrapper[4904]: I0214 11:15:37.901936 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d910fb92-a40a-4167-be6f-113f22c206b2-serving-cert\") pod \"d910fb92-a40a-4167-be6f-113f22c206b2\" (UID: \"d910fb92-a40a-4167-be6f-113f22c206b2\") " Feb 14 11:15:37 crc kubenswrapper[4904]: I0214 11:15:37.902079 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwngl\" (UniqueName: \"kubernetes.io/projected/d910fb92-a40a-4167-be6f-113f22c206b2-kube-api-access-jwngl\") pod \"d910fb92-a40a-4167-be6f-113f22c206b2\" (UID: \"d910fb92-a40a-4167-be6f-113f22c206b2\") " Feb 14 11:15:37 crc kubenswrapper[4904]: I0214 11:15:37.902115 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d910fb92-a40a-4167-be6f-113f22c206b2-config\") pod \"d910fb92-a40a-4167-be6f-113f22c206b2\" (UID: \"d910fb92-a40a-4167-be6f-113f22c206b2\") " Feb 14 11:15:37 crc kubenswrapper[4904]: I0214 11:15:37.902150 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d910fb92-a40a-4167-be6f-113f22c206b2-client-ca\") pod \"d910fb92-a40a-4167-be6f-113f22c206b2\" (UID: \"d910fb92-a40a-4167-be6f-113f22c206b2\") " Feb 14 11:15:37 crc kubenswrapper[4904]: I0214 11:15:37.903259 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d910fb92-a40a-4167-be6f-113f22c206b2-client-ca" (OuterVolumeSpecName: "client-ca") pod "d910fb92-a40a-4167-be6f-113f22c206b2" (UID: "d910fb92-a40a-4167-be6f-113f22c206b2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:15:37 crc kubenswrapper[4904]: I0214 11:15:37.903856 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d910fb92-a40a-4167-be6f-113f22c206b2-config" (OuterVolumeSpecName: "config") pod "d910fb92-a40a-4167-be6f-113f22c206b2" (UID: "d910fb92-a40a-4167-be6f-113f22c206b2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:15:37 crc kubenswrapper[4904]: I0214 11:15:37.908220 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d910fb92-a40a-4167-be6f-113f22c206b2-kube-api-access-jwngl" (OuterVolumeSpecName: "kube-api-access-jwngl") pod "d910fb92-a40a-4167-be6f-113f22c206b2" (UID: "d910fb92-a40a-4167-be6f-113f22c206b2"). InnerVolumeSpecName "kube-api-access-jwngl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:15:37 crc kubenswrapper[4904]: I0214 11:15:37.908250 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d910fb92-a40a-4167-be6f-113f22c206b2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d910fb92-a40a-4167-be6f-113f22c206b2" (UID: "d910fb92-a40a-4167-be6f-113f22c206b2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.003546 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a537ec4-ab86-4bc9-81d2-3870472b767a-config\") pod \"5a537ec4-ab86-4bc9-81d2-3870472b767a\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.003804 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5a537ec4-ab86-4bc9-81d2-3870472b767a-proxy-ca-bundles\") pod \"5a537ec4-ab86-4bc9-81d2-3870472b767a\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.003925 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpfx4\" (UniqueName: \"kubernetes.io/projected/5a537ec4-ab86-4bc9-81d2-3870472b767a-kube-api-access-dpfx4\") pod \"5a537ec4-ab86-4bc9-81d2-3870472b767a\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.004097 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a537ec4-ab86-4bc9-81d2-3870472b767a-serving-cert\") pod \"5a537ec4-ab86-4bc9-81d2-3870472b767a\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.004177 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5a537ec4-ab86-4bc9-81d2-3870472b767a-client-ca\") pod \"5a537ec4-ab86-4bc9-81d2-3870472b767a\" (UID: \"5a537ec4-ab86-4bc9-81d2-3870472b767a\") " Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.004418 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d910fb92-a40a-4167-be6f-113f22c206b2-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.004494 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwngl\" (UniqueName: \"kubernetes.io/projected/d910fb92-a40a-4167-be6f-113f22c206b2-kube-api-access-jwngl\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.004564 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d910fb92-a40a-4167-be6f-113f22c206b2-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.004637 4904 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d910fb92-a40a-4167-be6f-113f22c206b2-client-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.004455 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a537ec4-ab86-4bc9-81d2-3870472b767a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "5a537ec4-ab86-4bc9-81d2-3870472b767a" (UID: "5a537ec4-ab86-4bc9-81d2-3870472b767a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.004506 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a537ec4-ab86-4bc9-81d2-3870472b767a-config" (OuterVolumeSpecName: "config") pod "5a537ec4-ab86-4bc9-81d2-3870472b767a" (UID: "5a537ec4-ab86-4bc9-81d2-3870472b767a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.005296 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a537ec4-ab86-4bc9-81d2-3870472b767a-client-ca" (OuterVolumeSpecName: "client-ca") pod "5a537ec4-ab86-4bc9-81d2-3870472b767a" (UID: "5a537ec4-ab86-4bc9-81d2-3870472b767a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.006573 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a537ec4-ab86-4bc9-81d2-3870472b767a-kube-api-access-dpfx4" (OuterVolumeSpecName: "kube-api-access-dpfx4") pod "5a537ec4-ab86-4bc9-81d2-3870472b767a" (UID: "5a537ec4-ab86-4bc9-81d2-3870472b767a"). InnerVolumeSpecName "kube-api-access-dpfx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.007504 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a537ec4-ab86-4bc9-81d2-3870472b767a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5a537ec4-ab86-4bc9-81d2-3870472b767a" (UID: "5a537ec4-ab86-4bc9-81d2-3870472b767a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.105610 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a537ec4-ab86-4bc9-81d2-3870472b767a-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.105689 4904 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5a537ec4-ab86-4bc9-81d2-3870472b767a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.105706 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpfx4\" (UniqueName: \"kubernetes.io/projected/5a537ec4-ab86-4bc9-81d2-3870472b767a-kube-api-access-dpfx4\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.105718 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a537ec4-ab86-4bc9-81d2-3870472b767a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.105729 4904 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5a537ec4-ab86-4bc9-81d2-3870472b767a-client-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.259299 4904 generic.go:334] "Generic (PLEG): container finished" podID="d910fb92-a40a-4167-be6f-113f22c206b2" containerID="976a29dc369aa3b647f0f609fcfc3b5caccfdfc11a27fda35184577280c3470b" exitCode=0 Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.259361 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" event={"ID":"d910fb92-a40a-4167-be6f-113f22c206b2","Type":"ContainerDied","Data":"976a29dc369aa3b647f0f609fcfc3b5caccfdfc11a27fda35184577280c3470b"} Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.259387 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" event={"ID":"d910fb92-a40a-4167-be6f-113f22c206b2","Type":"ContainerDied","Data":"02efeac069d631dbfd5fb9698eb44b649260f643dc8eadeba4b881ddbcb02c75"} Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.259395 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.259419 4904 scope.go:117] "RemoveContainer" containerID="976a29dc369aa3b647f0f609fcfc3b5caccfdfc11a27fda35184577280c3470b" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.261478 4904 generic.go:334] "Generic (PLEG): container finished" podID="5a537ec4-ab86-4bc9-81d2-3870472b767a" containerID="9969e557a55d34107e04e866c201d4d8d5e8d7eb9f45fd44bde1c7338015a20e" exitCode=0 Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.261519 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" event={"ID":"5a537ec4-ab86-4bc9-81d2-3870472b767a","Type":"ContainerDied","Data":"9969e557a55d34107e04e866c201d4d8d5e8d7eb9f45fd44bde1c7338015a20e"} Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.261541 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.261549 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5lwz5" event={"ID":"5a537ec4-ab86-4bc9-81d2-3870472b767a","Type":"ContainerDied","Data":"d84f717518ef133711c1a0c3b66797a4af5047c5dfa24914b8f74a29d8e4d9cd"} Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.292152 4904 scope.go:117] "RemoveContainer" containerID="976a29dc369aa3b647f0f609fcfc3b5caccfdfc11a27fda35184577280c3470b" Feb 14 11:15:38 crc kubenswrapper[4904]: E0214 11:15:38.292610 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"976a29dc369aa3b647f0f609fcfc3b5caccfdfc11a27fda35184577280c3470b\": container with ID starting with 976a29dc369aa3b647f0f609fcfc3b5caccfdfc11a27fda35184577280c3470b not found: ID does not exist" containerID="976a29dc369aa3b647f0f609fcfc3b5caccfdfc11a27fda35184577280c3470b" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.292638 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"976a29dc369aa3b647f0f609fcfc3b5caccfdfc11a27fda35184577280c3470b"} err="failed to get container status \"976a29dc369aa3b647f0f609fcfc3b5caccfdfc11a27fda35184577280c3470b\": rpc error: code = NotFound desc = could not find container \"976a29dc369aa3b647f0f609fcfc3b5caccfdfc11a27fda35184577280c3470b\": container with ID starting with 976a29dc369aa3b647f0f609fcfc3b5caccfdfc11a27fda35184577280c3470b not found: ID does not exist" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.292657 4904 scope.go:117] "RemoveContainer" containerID="9969e557a55d34107e04e866c201d4d8d5e8d7eb9f45fd44bde1c7338015a20e" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.333143 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52"] Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.338870 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnf52"] Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.352574 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5lwz5"] Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.352892 4904 scope.go:117] "RemoveContainer" containerID="9969e557a55d34107e04e866c201d4d8d5e8d7eb9f45fd44bde1c7338015a20e" Feb 14 11:15:38 crc kubenswrapper[4904]: E0214 11:15:38.353234 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9969e557a55d34107e04e866c201d4d8d5e8d7eb9f45fd44bde1c7338015a20e\": container with ID starting with 9969e557a55d34107e04e866c201d4d8d5e8d7eb9f45fd44bde1c7338015a20e not found: ID does not exist" containerID="9969e557a55d34107e04e866c201d4d8d5e8d7eb9f45fd44bde1c7338015a20e" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.353260 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9969e557a55d34107e04e866c201d4d8d5e8d7eb9f45fd44bde1c7338015a20e"} err="failed to get container status \"9969e557a55d34107e04e866c201d4d8d5e8d7eb9f45fd44bde1c7338015a20e\": rpc error: code = NotFound desc = could not find container \"9969e557a55d34107e04e866c201d4d8d5e8d7eb9f45fd44bde1c7338015a20e\": container with ID starting with 9969e557a55d34107e04e866c201d4d8d5e8d7eb9f45fd44bde1c7338015a20e not found: ID does not exist" Feb 14 11:15:38 crc kubenswrapper[4904]: I0214 11:15:38.363910 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5lwz5"] Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.233613 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6d7f596867-bszpd"] Feb 14 11:15:39 crc kubenswrapper[4904]: E0214 11:15:39.235476 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d910fb92-a40a-4167-be6f-113f22c206b2" containerName="route-controller-manager" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.235492 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="d910fb92-a40a-4167-be6f-113f22c206b2" containerName="route-controller-manager" Feb 14 11:15:39 crc kubenswrapper[4904]: E0214 11:15:39.235522 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e31d6f5-239d-4418-b95d-dfe790798140" containerName="collect-profiles" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.235528 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e31d6f5-239d-4418-b95d-dfe790798140" containerName="collect-profiles" Feb 14 11:15:39 crc kubenswrapper[4904]: E0214 11:15:39.235539 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a537ec4-ab86-4bc9-81d2-3870472b767a" containerName="controller-manager" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.235545 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a537ec4-ab86-4bc9-81d2-3870472b767a" containerName="controller-manager" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.235708 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a537ec4-ab86-4bc9-81d2-3870472b767a" containerName="controller-manager" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.235724 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e31d6f5-239d-4418-b95d-dfe790798140" containerName="collect-profiles" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.235731 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="d910fb92-a40a-4167-be6f-113f22c206b2" containerName="route-controller-manager" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.239004 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t"] Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.239546 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.240015 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.246219 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.246376 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.246220 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.246518 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.246756 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.247753 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.247987 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.248188 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.248291 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.248495 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.248606 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.267631 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.267873 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d7f596867-bszpd"] Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.270661 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t"] Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.273593 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.420327 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8072b89e-7c97-4234-9d7b-74f8297358b7-serving-cert\") pod \"controller-manager-6d7f596867-bszpd\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.420374 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8072b89e-7c97-4234-9d7b-74f8297358b7-client-ca\") pod \"controller-manager-6d7f596867-bszpd\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.420534 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-serving-cert\") pod \"route-controller-manager-6d4598f449-7zp8t\" (UID: \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\") " pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.420636 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdqwh\" (UniqueName: \"kubernetes.io/projected/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-kube-api-access-tdqwh\") pod \"route-controller-manager-6d4598f449-7zp8t\" (UID: \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\") " pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.420736 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8072b89e-7c97-4234-9d7b-74f8297358b7-config\") pod \"controller-manager-6d7f596867-bszpd\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.420801 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8072b89e-7c97-4234-9d7b-74f8297358b7-proxy-ca-bundles\") pod \"controller-manager-6d7f596867-bszpd\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.420870 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v66fr\" (UniqueName: \"kubernetes.io/projected/8072b89e-7c97-4234-9d7b-74f8297358b7-kube-api-access-v66fr\") pod \"controller-manager-6d7f596867-bszpd\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.420910 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-config\") pod \"route-controller-manager-6d4598f449-7zp8t\" (UID: \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\") " pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.420938 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-client-ca\") pod \"route-controller-manager-6d4598f449-7zp8t\" (UID: \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\") " pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.522547 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-serving-cert\") pod \"route-controller-manager-6d4598f449-7zp8t\" (UID: \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\") " pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.522606 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdqwh\" (UniqueName: \"kubernetes.io/projected/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-kube-api-access-tdqwh\") pod \"route-controller-manager-6d4598f449-7zp8t\" (UID: \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\") " pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.522662 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8072b89e-7c97-4234-9d7b-74f8297358b7-config\") pod \"controller-manager-6d7f596867-bszpd\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.522990 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8072b89e-7c97-4234-9d7b-74f8297358b7-proxy-ca-bundles\") pod \"controller-manager-6d7f596867-bszpd\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.523980 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8072b89e-7c97-4234-9d7b-74f8297358b7-config\") pod \"controller-manager-6d7f596867-bszpd\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.524000 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8072b89e-7c97-4234-9d7b-74f8297358b7-proxy-ca-bundles\") pod \"controller-manager-6d7f596867-bszpd\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.523023 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v66fr\" (UniqueName: \"kubernetes.io/projected/8072b89e-7c97-4234-9d7b-74f8297358b7-kube-api-access-v66fr\") pod \"controller-manager-6d7f596867-bszpd\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.524074 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-client-ca\") pod \"route-controller-manager-6d4598f449-7zp8t\" (UID: \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\") " pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.524097 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-config\") pod \"route-controller-manager-6d4598f449-7zp8t\" (UID: \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\") " pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.524174 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8072b89e-7c97-4234-9d7b-74f8297358b7-serving-cert\") pod \"controller-manager-6d7f596867-bszpd\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.524276 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8072b89e-7c97-4234-9d7b-74f8297358b7-client-ca\") pod \"controller-manager-6d7f596867-bszpd\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.524760 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-client-ca\") pod \"route-controller-manager-6d4598f449-7zp8t\" (UID: \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\") " pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.525401 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8072b89e-7c97-4234-9d7b-74f8297358b7-client-ca\") pod \"controller-manager-6d7f596867-bszpd\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.525314 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-config\") pod \"route-controller-manager-6d4598f449-7zp8t\" (UID: \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\") " pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.526661 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-serving-cert\") pod \"route-controller-manager-6d4598f449-7zp8t\" (UID: \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\") " pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.528208 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8072b89e-7c97-4234-9d7b-74f8297358b7-serving-cert\") pod \"controller-manager-6d7f596867-bszpd\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.539051 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v66fr\" (UniqueName: \"kubernetes.io/projected/8072b89e-7c97-4234-9d7b-74f8297358b7-kube-api-access-v66fr\") pod \"controller-manager-6d7f596867-bszpd\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.541615 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdqwh\" (UniqueName: \"kubernetes.io/projected/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-kube-api-access-tdqwh\") pod \"route-controller-manager-6d4598f449-7zp8t\" (UID: \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\") " pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.586458 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.592704 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.813754 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d7f596867-bszpd"] Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.855955 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a537ec4-ab86-4bc9-81d2-3870472b767a" path="/var/lib/kubelet/pods/5a537ec4-ab86-4bc9-81d2-3870472b767a/volumes" Feb 14 11:15:39 crc kubenswrapper[4904]: I0214 11:15:39.859545 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d910fb92-a40a-4167-be6f-113f22c206b2" path="/var/lib/kubelet/pods/d910fb92-a40a-4167-be6f-113f22c206b2/volumes" Feb 14 11:15:40 crc kubenswrapper[4904]: I0214 11:15:40.053932 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t"] Feb 14 11:15:40 crc kubenswrapper[4904]: W0214 11:15:40.060100 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1778cb5d_f6bd_45a6_8e6d_6d45ab567a3c.slice/crio-3ec2390d97c10aa8908386e8a5b499431c504625d2ef728d4a8c1d64a1a3edfd WatchSource:0}: Error finding container 3ec2390d97c10aa8908386e8a5b499431c504625d2ef728d4a8c1d64a1a3edfd: Status 404 returned error can't find the container with id 3ec2390d97c10aa8908386e8a5b499431c504625d2ef728d4a8c1d64a1a3edfd Feb 14 11:15:40 crc kubenswrapper[4904]: I0214 11:15:40.284076 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" event={"ID":"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c","Type":"ContainerStarted","Data":"100127ecf0b73739ce32e85c347495248316660e4c6012d99e4073bc1e027d38"} Feb 14 11:15:40 crc kubenswrapper[4904]: I0214 11:15:40.284114 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" event={"ID":"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c","Type":"ContainerStarted","Data":"3ec2390d97c10aa8908386e8a5b499431c504625d2ef728d4a8c1d64a1a3edfd"} Feb 14 11:15:40 crc kubenswrapper[4904]: I0214 11:15:40.285080 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" Feb 14 11:15:40 crc kubenswrapper[4904]: I0214 11:15:40.286403 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" event={"ID":"8072b89e-7c97-4234-9d7b-74f8297358b7","Type":"ContainerStarted","Data":"cdce8bd59edbd608e5082a2d2c9f4cf44fdac8129ab0f9e10f03b20e9b31af1f"} Feb 14 11:15:40 crc kubenswrapper[4904]: I0214 11:15:40.286429 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" event={"ID":"8072b89e-7c97-4234-9d7b-74f8297358b7","Type":"ContainerStarted","Data":"f81f387e828d082c9e0231e842a959df76d5f93de55bd6532946d02e4c0e598b"} Feb 14 11:15:40 crc kubenswrapper[4904]: I0214 11:15:40.286537 4904 patch_prober.go:28] interesting pod/route-controller-manager-6d4598f449-7zp8t container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.65:8443/healthz\": dial tcp 10.217.0.65:8443: connect: connection refused" start-of-body= Feb 14 11:15:40 crc kubenswrapper[4904]: I0214 11:15:40.286567 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" podUID="1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.65:8443/healthz\": dial tcp 10.217.0.65:8443: connect: connection refused" Feb 14 11:15:40 crc kubenswrapper[4904]: I0214 11:15:40.286713 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:40 crc kubenswrapper[4904]: I0214 11:15:40.307982 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:40 crc kubenswrapper[4904]: I0214 11:15:40.334645 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" podStartSLOduration=3.334626285 podStartE2EDuration="3.334626285s" podCreationTimestamp="2026-02-14 11:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:15:40.334532801 +0000 UTC m=+331.147297462" watchObservedRunningTime="2026-02-14 11:15:40.334626285 +0000 UTC m=+331.147390946" Feb 14 11:15:40 crc kubenswrapper[4904]: I0214 11:15:40.336421 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" podStartSLOduration=3.336413516 podStartE2EDuration="3.336413516s" podCreationTimestamp="2026-02-14 11:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:15:40.31447871 +0000 UTC m=+331.127243371" watchObservedRunningTime="2026-02-14 11:15:40.336413516 +0000 UTC m=+331.149178177" Feb 14 11:15:41 crc kubenswrapper[4904]: I0214 11:15:41.309375 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" Feb 14 11:15:41 crc kubenswrapper[4904]: I0214 11:15:41.413804 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6d7f596867-bszpd"] Feb 14 11:15:41 crc kubenswrapper[4904]: I0214 11:15:41.435547 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t"] Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.314604 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" podUID="8072b89e-7c97-4234-9d7b-74f8297358b7" containerName="controller-manager" containerID="cri-o://cdce8bd59edbd608e5082a2d2c9f4cf44fdac8129ab0f9e10f03b20e9b31af1f" gracePeriod=30 Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.315231 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" podUID="1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c" containerName="route-controller-manager" containerID="cri-o://100127ecf0b73739ce32e85c347495248316660e4c6012d99e4073bc1e027d38" gracePeriod=30 Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.756268 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.761221 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.782674 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-648659b994-t9vjx"] Feb 14 11:15:43 crc kubenswrapper[4904]: E0214 11:15:43.782918 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8072b89e-7c97-4234-9d7b-74f8297358b7" containerName="controller-manager" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.782935 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="8072b89e-7c97-4234-9d7b-74f8297358b7" containerName="controller-manager" Feb 14 11:15:43 crc kubenswrapper[4904]: E0214 11:15:43.782945 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c" containerName="route-controller-manager" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.782952 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c" containerName="route-controller-manager" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.783048 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="8072b89e-7c97-4234-9d7b-74f8297358b7" containerName="controller-manager" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.783061 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c" containerName="route-controller-manager" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.783467 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.815352 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-648659b994-t9vjx"] Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.934606 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8072b89e-7c97-4234-9d7b-74f8297358b7-serving-cert\") pod \"8072b89e-7c97-4234-9d7b-74f8297358b7\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.934665 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-config\") pod \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\" (UID: \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\") " Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.934704 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v66fr\" (UniqueName: \"kubernetes.io/projected/8072b89e-7c97-4234-9d7b-74f8297358b7-kube-api-access-v66fr\") pod \"8072b89e-7c97-4234-9d7b-74f8297358b7\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.934756 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdqwh\" (UniqueName: \"kubernetes.io/projected/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-kube-api-access-tdqwh\") pod \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\" (UID: \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\") " Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.934801 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-client-ca\") pod \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\" (UID: \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\") " Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.934816 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-serving-cert\") pod \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\" (UID: \"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c\") " Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.934853 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8072b89e-7c97-4234-9d7b-74f8297358b7-proxy-ca-bundles\") pod \"8072b89e-7c97-4234-9d7b-74f8297358b7\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.934878 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8072b89e-7c97-4234-9d7b-74f8297358b7-client-ca\") pod \"8072b89e-7c97-4234-9d7b-74f8297358b7\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.934919 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8072b89e-7c97-4234-9d7b-74f8297358b7-config\") pod \"8072b89e-7c97-4234-9d7b-74f8297358b7\" (UID: \"8072b89e-7c97-4234-9d7b-74f8297358b7\") " Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.935061 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hpz9\" (UniqueName: \"kubernetes.io/projected/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-kube-api-access-9hpz9\") pod \"controller-manager-648659b994-t9vjx\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.935094 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-proxy-ca-bundles\") pod \"controller-manager-648659b994-t9vjx\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.935111 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-config\") pod \"controller-manager-648659b994-t9vjx\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.935135 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-serving-cert\") pod \"controller-manager-648659b994-t9vjx\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.935153 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-client-ca\") pod \"controller-manager-648659b994-t9vjx\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.935453 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-config" (OuterVolumeSpecName: "config") pod "1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c" (UID: "1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.935492 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-client-ca" (OuterVolumeSpecName: "client-ca") pod "1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c" (UID: "1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.935828 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8072b89e-7c97-4234-9d7b-74f8297358b7-client-ca" (OuterVolumeSpecName: "client-ca") pod "8072b89e-7c97-4234-9d7b-74f8297358b7" (UID: "8072b89e-7c97-4234-9d7b-74f8297358b7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.935980 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8072b89e-7c97-4234-9d7b-74f8297358b7-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "8072b89e-7c97-4234-9d7b-74f8297358b7" (UID: "8072b89e-7c97-4234-9d7b-74f8297358b7"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.936356 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8072b89e-7c97-4234-9d7b-74f8297358b7-config" (OuterVolumeSpecName: "config") pod "8072b89e-7c97-4234-9d7b-74f8297358b7" (UID: "8072b89e-7c97-4234-9d7b-74f8297358b7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.939989 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8072b89e-7c97-4234-9d7b-74f8297358b7-kube-api-access-v66fr" (OuterVolumeSpecName: "kube-api-access-v66fr") pod "8072b89e-7c97-4234-9d7b-74f8297358b7" (UID: "8072b89e-7c97-4234-9d7b-74f8297358b7"). InnerVolumeSpecName "kube-api-access-v66fr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.940120 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c" (UID: "1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.940554 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-kube-api-access-tdqwh" (OuterVolumeSpecName: "kube-api-access-tdqwh") pod "1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c" (UID: "1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c"). InnerVolumeSpecName "kube-api-access-tdqwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:15:43 crc kubenswrapper[4904]: I0214 11:15:43.948528 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8072b89e-7c97-4234-9d7b-74f8297358b7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8072b89e-7c97-4234-9d7b-74f8297358b7" (UID: "8072b89e-7c97-4234-9d7b-74f8297358b7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.035762 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-client-ca\") pod \"controller-manager-648659b994-t9vjx\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.035869 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hpz9\" (UniqueName: \"kubernetes.io/projected/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-kube-api-access-9hpz9\") pod \"controller-manager-648659b994-t9vjx\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.035902 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-proxy-ca-bundles\") pod \"controller-manager-648659b994-t9vjx\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.035920 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-config\") pod \"controller-manager-648659b994-t9vjx\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.035945 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-serving-cert\") pod \"controller-manager-648659b994-t9vjx\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.035981 4904 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-client-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.035994 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.036006 4904 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8072b89e-7c97-4234-9d7b-74f8297358b7-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.036017 4904 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8072b89e-7c97-4234-9d7b-74f8297358b7-client-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.036030 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8072b89e-7c97-4234-9d7b-74f8297358b7-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.036040 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8072b89e-7c97-4234-9d7b-74f8297358b7-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.036051 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.036061 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v66fr\" (UniqueName: \"kubernetes.io/projected/8072b89e-7c97-4234-9d7b-74f8297358b7-kube-api-access-v66fr\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.036073 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdqwh\" (UniqueName: \"kubernetes.io/projected/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c-kube-api-access-tdqwh\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.036852 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-client-ca\") pod \"controller-manager-648659b994-t9vjx\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.037055 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-proxy-ca-bundles\") pod \"controller-manager-648659b994-t9vjx\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.038332 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-config\") pod \"controller-manager-648659b994-t9vjx\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.039739 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-serving-cert\") pod \"controller-manager-648659b994-t9vjx\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.053329 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hpz9\" (UniqueName: \"kubernetes.io/projected/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-kube-api-access-9hpz9\") pod \"controller-manager-648659b994-t9vjx\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.099673 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.293281 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-648659b994-t9vjx"] Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.319940 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" event={"ID":"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475","Type":"ContainerStarted","Data":"e3049039930e9ce226e74b24e3c0d65b0e744d19b095e50bfb0f2a89692bab42"} Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.321111 4904 generic.go:334] "Generic (PLEG): container finished" podID="1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c" containerID="100127ecf0b73739ce32e85c347495248316660e4c6012d99e4073bc1e027d38" exitCode=0 Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.321184 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" event={"ID":"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c","Type":"ContainerDied","Data":"100127ecf0b73739ce32e85c347495248316660e4c6012d99e4073bc1e027d38"} Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.321203 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" event={"ID":"1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c","Type":"ContainerDied","Data":"3ec2390d97c10aa8908386e8a5b499431c504625d2ef728d4a8c1d64a1a3edfd"} Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.321219 4904 scope.go:117] "RemoveContainer" containerID="100127ecf0b73739ce32e85c347495248316660e4c6012d99e4073bc1e027d38" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.321322 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.323999 4904 generic.go:334] "Generic (PLEG): container finished" podID="8072b89e-7c97-4234-9d7b-74f8297358b7" containerID="cdce8bd59edbd608e5082a2d2c9f4cf44fdac8129ab0f9e10f03b20e9b31af1f" exitCode=0 Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.324055 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" event={"ID":"8072b89e-7c97-4234-9d7b-74f8297358b7","Type":"ContainerDied","Data":"cdce8bd59edbd608e5082a2d2c9f4cf44fdac8129ab0f9e10f03b20e9b31af1f"} Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.324078 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" event={"ID":"8072b89e-7c97-4234-9d7b-74f8297358b7","Type":"ContainerDied","Data":"f81f387e828d082c9e0231e842a959df76d5f93de55bd6532946d02e4c0e598b"} Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.324055 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d7f596867-bszpd" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.364223 4904 scope.go:117] "RemoveContainer" containerID="100127ecf0b73739ce32e85c347495248316660e4c6012d99e4073bc1e027d38" Feb 14 11:15:44 crc kubenswrapper[4904]: E0214 11:15:44.364579 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"100127ecf0b73739ce32e85c347495248316660e4c6012d99e4073bc1e027d38\": container with ID starting with 100127ecf0b73739ce32e85c347495248316660e4c6012d99e4073bc1e027d38 not found: ID does not exist" containerID="100127ecf0b73739ce32e85c347495248316660e4c6012d99e4073bc1e027d38" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.364609 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"100127ecf0b73739ce32e85c347495248316660e4c6012d99e4073bc1e027d38"} err="failed to get container status \"100127ecf0b73739ce32e85c347495248316660e4c6012d99e4073bc1e027d38\": rpc error: code = NotFound desc = could not find container \"100127ecf0b73739ce32e85c347495248316660e4c6012d99e4073bc1e027d38\": container with ID starting with 100127ecf0b73739ce32e85c347495248316660e4c6012d99e4073bc1e027d38 not found: ID does not exist" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.364627 4904 scope.go:117] "RemoveContainer" containerID="cdce8bd59edbd608e5082a2d2c9f4cf44fdac8129ab0f9e10f03b20e9b31af1f" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.383945 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t"] Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.385281 4904 scope.go:117] "RemoveContainer" containerID="cdce8bd59edbd608e5082a2d2c9f4cf44fdac8129ab0f9e10f03b20e9b31af1f" Feb 14 11:15:44 crc kubenswrapper[4904]: E0214 11:15:44.385805 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdce8bd59edbd608e5082a2d2c9f4cf44fdac8129ab0f9e10f03b20e9b31af1f\": container with ID starting with cdce8bd59edbd608e5082a2d2c9f4cf44fdac8129ab0f9e10f03b20e9b31af1f not found: ID does not exist" containerID="cdce8bd59edbd608e5082a2d2c9f4cf44fdac8129ab0f9e10f03b20e9b31af1f" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.385861 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdce8bd59edbd608e5082a2d2c9f4cf44fdac8129ab0f9e10f03b20e9b31af1f"} err="failed to get container status \"cdce8bd59edbd608e5082a2d2c9f4cf44fdac8129ab0f9e10f03b20e9b31af1f\": rpc error: code = NotFound desc = could not find container \"cdce8bd59edbd608e5082a2d2c9f4cf44fdac8129ab0f9e10f03b20e9b31af1f\": container with ID starting with cdce8bd59edbd608e5082a2d2c9f4cf44fdac8129ab0f9e10f03b20e9b31af1f not found: ID does not exist" Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.402458 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d4598f449-7zp8t"] Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.407275 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6d7f596867-bszpd"] Feb 14 11:15:44 crc kubenswrapper[4904]: I0214 11:15:44.410783 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6d7f596867-bszpd"] Feb 14 11:15:45 crc kubenswrapper[4904]: I0214 11:15:45.329870 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" event={"ID":"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475","Type":"ContainerStarted","Data":"1c5d3495483b77f9b172643439c6103cd477038c245f699e8e01d3a404e89c87"} Feb 14 11:15:45 crc kubenswrapper[4904]: I0214 11:15:45.330219 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:15:45 crc kubenswrapper[4904]: I0214 11:15:45.335431 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:15:45 crc kubenswrapper[4904]: I0214 11:15:45.349489 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" podStartSLOduration=4.349474051 podStartE2EDuration="4.349474051s" podCreationTimestamp="2026-02-14 11:15:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:15:45.34710595 +0000 UTC m=+336.159870611" watchObservedRunningTime="2026-02-14 11:15:45.349474051 +0000 UTC m=+336.162238712" Feb 14 11:15:45 crc kubenswrapper[4904]: I0214 11:15:45.843332 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c" path="/var/lib/kubelet/pods/1778cb5d-f6bd-45a6-8e6d-6d45ab567a3c/volumes" Feb 14 11:15:45 crc kubenswrapper[4904]: I0214 11:15:45.844069 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8072b89e-7c97-4234-9d7b-74f8297358b7" path="/var/lib/kubelet/pods/8072b89e-7c97-4234-9d7b-74f8297358b7/volumes" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.230413 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48"] Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.231087 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.233272 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.233785 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.233982 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.234099 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.234369 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.234480 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.242035 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48"] Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.363850 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7c6d254-0386-49f7-ab27-b4ac8401ac84-serving-cert\") pod \"route-controller-manager-6cddb56ddd-pln48\" (UID: \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\") " pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.363914 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f7c6d254-0386-49f7-ab27-b4ac8401ac84-client-ca\") pod \"route-controller-manager-6cddb56ddd-pln48\" (UID: \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\") " pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.364011 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7c6d254-0386-49f7-ab27-b4ac8401ac84-config\") pod \"route-controller-manager-6cddb56ddd-pln48\" (UID: \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\") " pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.364046 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzd26\" (UniqueName: \"kubernetes.io/projected/f7c6d254-0386-49f7-ab27-b4ac8401ac84-kube-api-access-wzd26\") pod \"route-controller-manager-6cddb56ddd-pln48\" (UID: \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\") " pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.465534 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f7c6d254-0386-49f7-ab27-b4ac8401ac84-client-ca\") pod \"route-controller-manager-6cddb56ddd-pln48\" (UID: \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\") " pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.465765 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7c6d254-0386-49f7-ab27-b4ac8401ac84-config\") pod \"route-controller-manager-6cddb56ddd-pln48\" (UID: \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\") " pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.465868 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzd26\" (UniqueName: \"kubernetes.io/projected/f7c6d254-0386-49f7-ab27-b4ac8401ac84-kube-api-access-wzd26\") pod \"route-controller-manager-6cddb56ddd-pln48\" (UID: \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\") " pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.465965 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7c6d254-0386-49f7-ab27-b4ac8401ac84-serving-cert\") pod \"route-controller-manager-6cddb56ddd-pln48\" (UID: \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\") " pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.466924 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7c6d254-0386-49f7-ab27-b4ac8401ac84-config\") pod \"route-controller-manager-6cddb56ddd-pln48\" (UID: \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\") " pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.467494 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f7c6d254-0386-49f7-ab27-b4ac8401ac84-client-ca\") pod \"route-controller-manager-6cddb56ddd-pln48\" (UID: \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\") " pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.470997 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7c6d254-0386-49f7-ab27-b4ac8401ac84-serving-cert\") pod \"route-controller-manager-6cddb56ddd-pln48\" (UID: \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\") " pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.480357 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzd26\" (UniqueName: \"kubernetes.io/projected/f7c6d254-0386-49f7-ab27-b4ac8401ac84-kube-api-access-wzd26\") pod \"route-controller-manager-6cddb56ddd-pln48\" (UID: \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\") " pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.544400 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" Feb 14 11:15:46 crc kubenswrapper[4904]: I0214 11:15:46.945325 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48"] Feb 14 11:15:47 crc kubenswrapper[4904]: I0214 11:15:47.342700 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" event={"ID":"f7c6d254-0386-49f7-ab27-b4ac8401ac84","Type":"ContainerStarted","Data":"b6e8d5a60a26f33d029d5b79fd52476a4633afdc9d158abb4797ee304b84d4a0"} Feb 14 11:15:47 crc kubenswrapper[4904]: I0214 11:15:47.342738 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" event={"ID":"f7c6d254-0386-49f7-ab27-b4ac8401ac84","Type":"ContainerStarted","Data":"5db9f52051b4cd895b9afcc0af7d74c43852c6ea840a2cce6445696c0a39bc75"} Feb 14 11:15:47 crc kubenswrapper[4904]: I0214 11:15:47.342942 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" Feb 14 11:15:47 crc kubenswrapper[4904]: I0214 11:15:47.349893 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" Feb 14 11:15:47 crc kubenswrapper[4904]: I0214 11:15:47.363443 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" podStartSLOduration=6.363427109 podStartE2EDuration="6.363427109s" podCreationTimestamp="2026-02-14 11:15:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:15:47.361323718 +0000 UTC m=+338.174088379" watchObservedRunningTime="2026-02-14 11:15:47.363427109 +0000 UTC m=+338.176191770" Feb 14 11:15:57 crc kubenswrapper[4904]: I0214 11:15:57.341185 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48"] Feb 14 11:15:57 crc kubenswrapper[4904]: I0214 11:15:57.341882 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" podUID="f7c6d254-0386-49f7-ab27-b4ac8401ac84" containerName="route-controller-manager" containerID="cri-o://b6e8d5a60a26f33d029d5b79fd52476a4633afdc9d158abb4797ee304b84d4a0" gracePeriod=30 Feb 14 11:15:57 crc kubenswrapper[4904]: I0214 11:15:57.789223 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" Feb 14 11:15:57 crc kubenswrapper[4904]: I0214 11:15:57.804355 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7c6d254-0386-49f7-ab27-b4ac8401ac84-config\") pod \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\" (UID: \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\") " Feb 14 11:15:57 crc kubenswrapper[4904]: I0214 11:15:57.805522 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7c6d254-0386-49f7-ab27-b4ac8401ac84-config" (OuterVolumeSpecName: "config") pod "f7c6d254-0386-49f7-ab27-b4ac8401ac84" (UID: "f7c6d254-0386-49f7-ab27-b4ac8401ac84"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:15:57 crc kubenswrapper[4904]: I0214 11:15:57.905538 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f7c6d254-0386-49f7-ab27-b4ac8401ac84-client-ca\") pod \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\" (UID: \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\") " Feb 14 11:15:57 crc kubenswrapper[4904]: I0214 11:15:57.905622 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzd26\" (UniqueName: \"kubernetes.io/projected/f7c6d254-0386-49f7-ab27-b4ac8401ac84-kube-api-access-wzd26\") pod \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\" (UID: \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\") " Feb 14 11:15:57 crc kubenswrapper[4904]: I0214 11:15:57.905643 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7c6d254-0386-49f7-ab27-b4ac8401ac84-serving-cert\") pod \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\" (UID: \"f7c6d254-0386-49f7-ab27-b4ac8401ac84\") " Feb 14 11:15:57 crc kubenswrapper[4904]: I0214 11:15:57.905863 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7c6d254-0386-49f7-ab27-b4ac8401ac84-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:57 crc kubenswrapper[4904]: I0214 11:15:57.906062 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7c6d254-0386-49f7-ab27-b4ac8401ac84-client-ca" (OuterVolumeSpecName: "client-ca") pod "f7c6d254-0386-49f7-ab27-b4ac8401ac84" (UID: "f7c6d254-0386-49f7-ab27-b4ac8401ac84"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:15:57 crc kubenswrapper[4904]: I0214 11:15:57.910904 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7c6d254-0386-49f7-ab27-b4ac8401ac84-kube-api-access-wzd26" (OuterVolumeSpecName: "kube-api-access-wzd26") pod "f7c6d254-0386-49f7-ab27-b4ac8401ac84" (UID: "f7c6d254-0386-49f7-ab27-b4ac8401ac84"). InnerVolumeSpecName "kube-api-access-wzd26". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:15:57 crc kubenswrapper[4904]: I0214 11:15:57.911417 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7c6d254-0386-49f7-ab27-b4ac8401ac84-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f7c6d254-0386-49f7-ab27-b4ac8401ac84" (UID: "f7c6d254-0386-49f7-ab27-b4ac8401ac84"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:15:58 crc kubenswrapper[4904]: I0214 11:15:58.008079 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzd26\" (UniqueName: \"kubernetes.io/projected/f7c6d254-0386-49f7-ab27-b4ac8401ac84-kube-api-access-wzd26\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:58 crc kubenswrapper[4904]: I0214 11:15:58.008134 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7c6d254-0386-49f7-ab27-b4ac8401ac84-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:58 crc kubenswrapper[4904]: I0214 11:15:58.008144 4904 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f7c6d254-0386-49f7-ab27-b4ac8401ac84-client-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:15:58 crc kubenswrapper[4904]: I0214 11:15:58.398536 4904 generic.go:334] "Generic (PLEG): container finished" podID="f7c6d254-0386-49f7-ab27-b4ac8401ac84" containerID="b6e8d5a60a26f33d029d5b79fd52476a4633afdc9d158abb4797ee304b84d4a0" exitCode=0 Feb 14 11:15:58 crc kubenswrapper[4904]: I0214 11:15:58.398627 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" Feb 14 11:15:58 crc kubenswrapper[4904]: I0214 11:15:58.398666 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" event={"ID":"f7c6d254-0386-49f7-ab27-b4ac8401ac84","Type":"ContainerDied","Data":"b6e8d5a60a26f33d029d5b79fd52476a4633afdc9d158abb4797ee304b84d4a0"} Feb 14 11:15:58 crc kubenswrapper[4904]: I0214 11:15:58.401161 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48" event={"ID":"f7c6d254-0386-49f7-ab27-b4ac8401ac84","Type":"ContainerDied","Data":"5db9f52051b4cd895b9afcc0af7d74c43852c6ea840a2cce6445696c0a39bc75"} Feb 14 11:15:58 crc kubenswrapper[4904]: I0214 11:15:58.401200 4904 scope.go:117] "RemoveContainer" containerID="b6e8d5a60a26f33d029d5b79fd52476a4633afdc9d158abb4797ee304b84d4a0" Feb 14 11:15:58 crc kubenswrapper[4904]: I0214 11:15:58.422822 4904 scope.go:117] "RemoveContainer" containerID="b6e8d5a60a26f33d029d5b79fd52476a4633afdc9d158abb4797ee304b84d4a0" Feb 14 11:15:58 crc kubenswrapper[4904]: E0214 11:15:58.423553 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6e8d5a60a26f33d029d5b79fd52476a4633afdc9d158abb4797ee304b84d4a0\": container with ID starting with b6e8d5a60a26f33d029d5b79fd52476a4633afdc9d158abb4797ee304b84d4a0 not found: ID does not exist" containerID="b6e8d5a60a26f33d029d5b79fd52476a4633afdc9d158abb4797ee304b84d4a0" Feb 14 11:15:58 crc kubenswrapper[4904]: I0214 11:15:58.424693 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6e8d5a60a26f33d029d5b79fd52476a4633afdc9d158abb4797ee304b84d4a0"} err="failed to get container status \"b6e8d5a60a26f33d029d5b79fd52476a4633afdc9d158abb4797ee304b84d4a0\": rpc error: code = NotFound desc = could not find container \"b6e8d5a60a26f33d029d5b79fd52476a4633afdc9d158abb4797ee304b84d4a0\": container with ID starting with b6e8d5a60a26f33d029d5b79fd52476a4633afdc9d158abb4797ee304b84d4a0 not found: ID does not exist" Feb 14 11:15:58 crc kubenswrapper[4904]: I0214 11:15:58.454892 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48"] Feb 14 11:15:58 crc kubenswrapper[4904]: I0214 11:15:58.465409 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6cddb56ddd-pln48"] Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.250731 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2"] Feb 14 11:15:59 crc kubenswrapper[4904]: E0214 11:15:59.251208 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7c6d254-0386-49f7-ab27-b4ac8401ac84" containerName="route-controller-manager" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.251227 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7c6d254-0386-49f7-ab27-b4ac8401ac84" containerName="route-controller-manager" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.251476 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7c6d254-0386-49f7-ab27-b4ac8401ac84" containerName="route-controller-manager" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.252197 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.262050 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.262312 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.262102 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.262240 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.265627 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.265791 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.278041 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2"] Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.425965 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed04fa91-e6c6-40f2-9043-3fa1de5aef41-serving-cert\") pod \"route-controller-manager-89ddffbc6-w8sm2\" (UID: \"ed04fa91-e6c6-40f2-9043-3fa1de5aef41\") " pod="openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.426730 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed04fa91-e6c6-40f2-9043-3fa1de5aef41-client-ca\") pod \"route-controller-manager-89ddffbc6-w8sm2\" (UID: \"ed04fa91-e6c6-40f2-9043-3fa1de5aef41\") " pod="openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.426884 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed04fa91-e6c6-40f2-9043-3fa1de5aef41-config\") pod \"route-controller-manager-89ddffbc6-w8sm2\" (UID: \"ed04fa91-e6c6-40f2-9043-3fa1de5aef41\") " pod="openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.426995 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8lgm\" (UniqueName: \"kubernetes.io/projected/ed04fa91-e6c6-40f2-9043-3fa1de5aef41-kube-api-access-n8lgm\") pod \"route-controller-manager-89ddffbc6-w8sm2\" (UID: \"ed04fa91-e6c6-40f2-9043-3fa1de5aef41\") " pod="openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.527583 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed04fa91-e6c6-40f2-9043-3fa1de5aef41-client-ca\") pod \"route-controller-manager-89ddffbc6-w8sm2\" (UID: \"ed04fa91-e6c6-40f2-9043-3fa1de5aef41\") " pod="openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.527638 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed04fa91-e6c6-40f2-9043-3fa1de5aef41-config\") pod \"route-controller-manager-89ddffbc6-w8sm2\" (UID: \"ed04fa91-e6c6-40f2-9043-3fa1de5aef41\") " pod="openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.527673 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8lgm\" (UniqueName: \"kubernetes.io/projected/ed04fa91-e6c6-40f2-9043-3fa1de5aef41-kube-api-access-n8lgm\") pod \"route-controller-manager-89ddffbc6-w8sm2\" (UID: \"ed04fa91-e6c6-40f2-9043-3fa1de5aef41\") " pod="openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.527713 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed04fa91-e6c6-40f2-9043-3fa1de5aef41-serving-cert\") pod \"route-controller-manager-89ddffbc6-w8sm2\" (UID: \"ed04fa91-e6c6-40f2-9043-3fa1de5aef41\") " pod="openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.528767 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ed04fa91-e6c6-40f2-9043-3fa1de5aef41-client-ca\") pod \"route-controller-manager-89ddffbc6-w8sm2\" (UID: \"ed04fa91-e6c6-40f2-9043-3fa1de5aef41\") " pod="openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.529243 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed04fa91-e6c6-40f2-9043-3fa1de5aef41-config\") pod \"route-controller-manager-89ddffbc6-w8sm2\" (UID: \"ed04fa91-e6c6-40f2-9043-3fa1de5aef41\") " pod="openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.544751 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed04fa91-e6c6-40f2-9043-3fa1de5aef41-serving-cert\") pod \"route-controller-manager-89ddffbc6-w8sm2\" (UID: \"ed04fa91-e6c6-40f2-9043-3fa1de5aef41\") " pod="openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.546899 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8lgm\" (UniqueName: \"kubernetes.io/projected/ed04fa91-e6c6-40f2-9043-3fa1de5aef41-kube-api-access-n8lgm\") pod \"route-controller-manager-89ddffbc6-w8sm2\" (UID: \"ed04fa91-e6c6-40f2-9043-3fa1de5aef41\") " pod="openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.578612 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.845199 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7c6d254-0386-49f7-ab27-b4ac8401ac84" path="/var/lib/kubelet/pods/f7c6d254-0386-49f7-ab27-b4ac8401ac84/volumes" Feb 14 11:15:59 crc kubenswrapper[4904]: I0214 11:15:59.943117 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2"] Feb 14 11:15:59 crc kubenswrapper[4904]: W0214 11:15:59.949028 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded04fa91_e6c6_40f2_9043_3fa1de5aef41.slice/crio-a416293c32c69a765a38c3689518691f97107e93485a53ccaf8262161c4e0655 WatchSource:0}: Error finding container a416293c32c69a765a38c3689518691f97107e93485a53ccaf8262161c4e0655: Status 404 returned error can't find the container with id a416293c32c69a765a38c3689518691f97107e93485a53ccaf8262161c4e0655 Feb 14 11:16:00 crc kubenswrapper[4904]: I0214 11:16:00.416966 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2" event={"ID":"ed04fa91-e6c6-40f2-9043-3fa1de5aef41","Type":"ContainerStarted","Data":"aa7335cf37882c7ae8be183363a91c46dafd89d2edac249189f65ad4f953a922"} Feb 14 11:16:00 crc kubenswrapper[4904]: I0214 11:16:00.417013 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2" event={"ID":"ed04fa91-e6c6-40f2-9043-3fa1de5aef41","Type":"ContainerStarted","Data":"a416293c32c69a765a38c3689518691f97107e93485a53ccaf8262161c4e0655"} Feb 14 11:16:00 crc kubenswrapper[4904]: I0214 11:16:00.417410 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2" Feb 14 11:16:00 crc kubenswrapper[4904]: I0214 11:16:00.434924 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2" podStartSLOduration=3.434902679 podStartE2EDuration="3.434902679s" podCreationTimestamp="2026-02-14 11:15:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:16:00.432590781 +0000 UTC m=+351.245355462" watchObservedRunningTime="2026-02-14 11:16:00.434902679 +0000 UTC m=+351.247667350" Feb 14 11:16:00 crc kubenswrapper[4904]: I0214 11:16:00.662189 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-89ddffbc6-w8sm2" Feb 14 11:16:14 crc kubenswrapper[4904]: I0214 11:16:14.840091 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tdzwv"] Feb 14 11:16:14 crc kubenswrapper[4904]: I0214 11:16:14.841413 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:14 crc kubenswrapper[4904]: I0214 11:16:14.853956 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tdzwv"] Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.027341 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/293fd9e9-2486-4211-86a5-b055a8b82e08-registry-tls\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.027409 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/293fd9e9-2486-4211-86a5-b055a8b82e08-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.027444 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv9wz\" (UniqueName: \"kubernetes.io/projected/293fd9e9-2486-4211-86a5-b055a8b82e08-kube-api-access-bv9wz\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.027540 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/293fd9e9-2486-4211-86a5-b055a8b82e08-bound-sa-token\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.027576 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/293fd9e9-2486-4211-86a5-b055a8b82e08-trusted-ca\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.027616 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.027644 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/293fd9e9-2486-4211-86a5-b055a8b82e08-registry-certificates\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.027712 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/293fd9e9-2486-4211-86a5-b055a8b82e08-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.062089 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.129223 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/293fd9e9-2486-4211-86a5-b055a8b82e08-bound-sa-token\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.129794 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/293fd9e9-2486-4211-86a5-b055a8b82e08-trusted-ca\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.129954 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/293fd9e9-2486-4211-86a5-b055a8b82e08-registry-certificates\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.130125 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/293fd9e9-2486-4211-86a5-b055a8b82e08-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.130221 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/293fd9e9-2486-4211-86a5-b055a8b82e08-registry-tls\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.130299 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/293fd9e9-2486-4211-86a5-b055a8b82e08-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.130382 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv9wz\" (UniqueName: \"kubernetes.io/projected/293fd9e9-2486-4211-86a5-b055a8b82e08-kube-api-access-bv9wz\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.130502 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/293fd9e9-2486-4211-86a5-b055a8b82e08-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.131064 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/293fd9e9-2486-4211-86a5-b055a8b82e08-trusted-ca\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.131929 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/293fd9e9-2486-4211-86a5-b055a8b82e08-registry-certificates\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.135968 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/293fd9e9-2486-4211-86a5-b055a8b82e08-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.139385 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/293fd9e9-2486-4211-86a5-b055a8b82e08-registry-tls\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.150174 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv9wz\" (UniqueName: \"kubernetes.io/projected/293fd9e9-2486-4211-86a5-b055a8b82e08-kube-api-access-bv9wz\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.150455 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/293fd9e9-2486-4211-86a5-b055a8b82e08-bound-sa-token\") pod \"image-registry-66df7c8f76-tdzwv\" (UID: \"293fd9e9-2486-4211-86a5-b055a8b82e08\") " pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.156906 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:15 crc kubenswrapper[4904]: I0214 11:16:15.570534 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tdzwv"] Feb 14 11:16:16 crc kubenswrapper[4904]: I0214 11:16:16.382500 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:16:16 crc kubenswrapper[4904]: I0214 11:16:16.382561 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:16:16 crc kubenswrapper[4904]: I0214 11:16:16.492224 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" event={"ID":"293fd9e9-2486-4211-86a5-b055a8b82e08","Type":"ContainerStarted","Data":"2d4078125930ce69878492a5fa21b1a4b3dda118e2410d09429375978fb9b04e"} Feb 14 11:16:16 crc kubenswrapper[4904]: I0214 11:16:16.492266 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" event={"ID":"293fd9e9-2486-4211-86a5-b055a8b82e08","Type":"ContainerStarted","Data":"9943f15035d6a264fb82ce558068fe91a6e0aef2d5c561e4e9aadd377e242bf6"} Feb 14 11:16:16 crc kubenswrapper[4904]: I0214 11:16:16.493105 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:35 crc kubenswrapper[4904]: I0214 11:16:35.161184 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" Feb 14 11:16:35 crc kubenswrapper[4904]: I0214 11:16:35.181574 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-tdzwv" podStartSLOduration=21.181557358 podStartE2EDuration="21.181557358s" podCreationTimestamp="2026-02-14 11:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:16:16.508240589 +0000 UTC m=+367.321005270" watchObservedRunningTime="2026-02-14 11:16:35.181557358 +0000 UTC m=+385.994322019" Feb 14 11:16:35 crc kubenswrapper[4904]: I0214 11:16:35.214656 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zbh4h"] Feb 14 11:16:37 crc kubenswrapper[4904]: I0214 11:16:37.358371 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-648659b994-t9vjx"] Feb 14 11:16:37 crc kubenswrapper[4904]: I0214 11:16:37.358578 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" podUID="a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475" containerName="controller-manager" containerID="cri-o://1c5d3495483b77f9b172643439c6103cd477038c245f699e8e01d3a404e89c87" gracePeriod=30 Feb 14 11:16:37 crc kubenswrapper[4904]: I0214 11:16:37.606179 4904 generic.go:334] "Generic (PLEG): container finished" podID="a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475" containerID="1c5d3495483b77f9b172643439c6103cd477038c245f699e8e01d3a404e89c87" exitCode=0 Feb 14 11:16:37 crc kubenswrapper[4904]: I0214 11:16:37.606279 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" event={"ID":"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475","Type":"ContainerDied","Data":"1c5d3495483b77f9b172643439c6103cd477038c245f699e8e01d3a404e89c87"} Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.255786 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.421761 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-proxy-ca-bundles\") pod \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.421813 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-serving-cert\") pod \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.421989 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-config\") pod \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.422023 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hpz9\" (UniqueName: \"kubernetes.io/projected/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-kube-api-access-9hpz9\") pod \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.422066 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-client-ca\") pod \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\" (UID: \"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475\") " Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.422676 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-client-ca" (OuterVolumeSpecName: "client-ca") pod "a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475" (UID: "a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.422758 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-config" (OuterVolumeSpecName: "config") pod "a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475" (UID: "a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.422810 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475" (UID: "a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.427369 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-kube-api-access-9hpz9" (OuterVolumeSpecName: "kube-api-access-9hpz9") pod "a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475" (UID: "a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475"). InnerVolumeSpecName "kube-api-access-9hpz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.430239 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475" (UID: "a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.523227 4904 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-client-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.523265 4904 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.523276 4904 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.523304 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.523314 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hpz9\" (UniqueName: \"kubernetes.io/projected/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475-kube-api-access-9hpz9\") on node \"crc\" DevicePath \"\"" Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.611584 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" event={"ID":"a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475","Type":"ContainerDied","Data":"e3049039930e9ce226e74b24e3c0d65b0e744d19b095e50bfb0f2a89692bab42"} Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.611631 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-648659b994-t9vjx" Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.611636 4904 scope.go:117] "RemoveContainer" containerID="1c5d3495483b77f9b172643439c6103cd477038c245f699e8e01d3a404e89c87" Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.646809 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-648659b994-t9vjx"] Feb 14 11:16:38 crc kubenswrapper[4904]: I0214 11:16:38.651752 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-648659b994-t9vjx"] Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.262493 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5c76b5df59-msxl9"] Feb 14 11:16:39 crc kubenswrapper[4904]: E0214 11:16:39.262729 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475" containerName="controller-manager" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.262740 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475" containerName="controller-manager" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.262828 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475" containerName="controller-manager" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.263459 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.265959 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.266307 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.266317 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.269146 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.269147 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.274552 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.283432 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.285615 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5c76b5df59-msxl9"] Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.436786 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78833ebc-3a13-486d-81ad-9a10296219cb-config\") pod \"controller-manager-5c76b5df59-msxl9\" (UID: \"78833ebc-3a13-486d-81ad-9a10296219cb\") " pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.436863 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/78833ebc-3a13-486d-81ad-9a10296219cb-client-ca\") pod \"controller-manager-5c76b5df59-msxl9\" (UID: \"78833ebc-3a13-486d-81ad-9a10296219cb\") " pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.436937 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8jdd\" (UniqueName: \"kubernetes.io/projected/78833ebc-3a13-486d-81ad-9a10296219cb-kube-api-access-f8jdd\") pod \"controller-manager-5c76b5df59-msxl9\" (UID: \"78833ebc-3a13-486d-81ad-9a10296219cb\") " pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.436972 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78833ebc-3a13-486d-81ad-9a10296219cb-serving-cert\") pod \"controller-manager-5c76b5df59-msxl9\" (UID: \"78833ebc-3a13-486d-81ad-9a10296219cb\") " pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.436990 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/78833ebc-3a13-486d-81ad-9a10296219cb-proxy-ca-bundles\") pod \"controller-manager-5c76b5df59-msxl9\" (UID: \"78833ebc-3a13-486d-81ad-9a10296219cb\") " pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.538779 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78833ebc-3a13-486d-81ad-9a10296219cb-config\") pod \"controller-manager-5c76b5df59-msxl9\" (UID: \"78833ebc-3a13-486d-81ad-9a10296219cb\") " pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.541103 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/78833ebc-3a13-486d-81ad-9a10296219cb-client-ca\") pod \"controller-manager-5c76b5df59-msxl9\" (UID: \"78833ebc-3a13-486d-81ad-9a10296219cb\") " pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.541157 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8jdd\" (UniqueName: \"kubernetes.io/projected/78833ebc-3a13-486d-81ad-9a10296219cb-kube-api-access-f8jdd\") pod \"controller-manager-5c76b5df59-msxl9\" (UID: \"78833ebc-3a13-486d-81ad-9a10296219cb\") " pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.541180 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78833ebc-3a13-486d-81ad-9a10296219cb-serving-cert\") pod \"controller-manager-5c76b5df59-msxl9\" (UID: \"78833ebc-3a13-486d-81ad-9a10296219cb\") " pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.541198 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/78833ebc-3a13-486d-81ad-9a10296219cb-proxy-ca-bundles\") pod \"controller-manager-5c76b5df59-msxl9\" (UID: \"78833ebc-3a13-486d-81ad-9a10296219cb\") " pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.541335 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78833ebc-3a13-486d-81ad-9a10296219cb-config\") pod \"controller-manager-5c76b5df59-msxl9\" (UID: \"78833ebc-3a13-486d-81ad-9a10296219cb\") " pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.542084 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/78833ebc-3a13-486d-81ad-9a10296219cb-client-ca\") pod \"controller-manager-5c76b5df59-msxl9\" (UID: \"78833ebc-3a13-486d-81ad-9a10296219cb\") " pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.542237 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/78833ebc-3a13-486d-81ad-9a10296219cb-proxy-ca-bundles\") pod \"controller-manager-5c76b5df59-msxl9\" (UID: \"78833ebc-3a13-486d-81ad-9a10296219cb\") " pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.544496 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78833ebc-3a13-486d-81ad-9a10296219cb-serving-cert\") pod \"controller-manager-5c76b5df59-msxl9\" (UID: \"78833ebc-3a13-486d-81ad-9a10296219cb\") " pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.561337 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8jdd\" (UniqueName: \"kubernetes.io/projected/78833ebc-3a13-486d-81ad-9a10296219cb-kube-api-access-f8jdd\") pod \"controller-manager-5c76b5df59-msxl9\" (UID: \"78833ebc-3a13-486d-81ad-9a10296219cb\") " pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.583228 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.863486 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475" path="/var/lib/kubelet/pods/a4a47e4a-5d0f-4b8b-8e65-7a4d8578b475/volumes" Feb 14 11:16:39 crc kubenswrapper[4904]: I0214 11:16:39.988774 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5c76b5df59-msxl9"] Feb 14 11:16:40 crc kubenswrapper[4904]: I0214 11:16:40.628645 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" event={"ID":"78833ebc-3a13-486d-81ad-9a10296219cb","Type":"ContainerStarted","Data":"3ee58084182b3e7b8bbeab321b6155723497ce5246e9f0af38ce7e5e2e02aae1"} Feb 14 11:16:40 crc kubenswrapper[4904]: I0214 11:16:40.629632 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" event={"ID":"78833ebc-3a13-486d-81ad-9a10296219cb","Type":"ContainerStarted","Data":"0d5c7b7b86af7eaf849352674668d3d27c1e5674c6c89c2e654db384114cc5d4"} Feb 14 11:16:40 crc kubenswrapper[4904]: I0214 11:16:40.629733 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" Feb 14 11:16:40 crc kubenswrapper[4904]: I0214 11:16:40.634401 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" Feb 14 11:16:40 crc kubenswrapper[4904]: I0214 11:16:40.648941 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5c76b5df59-msxl9" podStartSLOduration=3.6489157150000002 podStartE2EDuration="3.648915715s" podCreationTimestamp="2026-02-14 11:16:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:16:40.643241362 +0000 UTC m=+391.456006043" watchObservedRunningTime="2026-02-14 11:16:40.648915715 +0000 UTC m=+391.461680386" Feb 14 11:16:46 crc kubenswrapper[4904]: I0214 11:16:46.383253 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:16:46 crc kubenswrapper[4904]: I0214 11:16:46.383793 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.247646 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" podUID="d97a9335-cf15-4b30-ba2e-846e0daed72d" containerName="registry" containerID="cri-o://a50f842bce824e3f5855d72f7e9101d5d8ecedc24956419d4ddc8f2ee345d29b" gracePeriod=30 Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.689135 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.743479 4904 generic.go:334] "Generic (PLEG): container finished" podID="d97a9335-cf15-4b30-ba2e-846e0daed72d" containerID="a50f842bce824e3f5855d72f7e9101d5d8ecedc24956419d4ddc8f2ee345d29b" exitCode=0 Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.743520 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" event={"ID":"d97a9335-cf15-4b30-ba2e-846e0daed72d","Type":"ContainerDied","Data":"a50f842bce824e3f5855d72f7e9101d5d8ecedc24956419d4ddc8f2ee345d29b"} Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.743545 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" event={"ID":"d97a9335-cf15-4b30-ba2e-846e0daed72d","Type":"ContainerDied","Data":"8d9cdb862c4416609468bf22f5a8fd04c4e1d268f029a75897a9b5cff99c922a"} Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.743553 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zbh4h" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.743561 4904 scope.go:117] "RemoveContainer" containerID="a50f842bce824e3f5855d72f7e9101d5d8ecedc24956419d4ddc8f2ee345d29b" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.761724 4904 scope.go:117] "RemoveContainer" containerID="a50f842bce824e3f5855d72f7e9101d5d8ecedc24956419d4ddc8f2ee345d29b" Feb 14 11:17:00 crc kubenswrapper[4904]: E0214 11:17:00.762296 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a50f842bce824e3f5855d72f7e9101d5d8ecedc24956419d4ddc8f2ee345d29b\": container with ID starting with a50f842bce824e3f5855d72f7e9101d5d8ecedc24956419d4ddc8f2ee345d29b not found: ID does not exist" containerID="a50f842bce824e3f5855d72f7e9101d5d8ecedc24956419d4ddc8f2ee345d29b" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.762336 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a50f842bce824e3f5855d72f7e9101d5d8ecedc24956419d4ddc8f2ee345d29b"} err="failed to get container status \"a50f842bce824e3f5855d72f7e9101d5d8ecedc24956419d4ddc8f2ee345d29b\": rpc error: code = NotFound desc = could not find container \"a50f842bce824e3f5855d72f7e9101d5d8ecedc24956419d4ddc8f2ee345d29b\": container with ID starting with a50f842bce824e3f5855d72f7e9101d5d8ecedc24956419d4ddc8f2ee345d29b not found: ID does not exist" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.871608 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d97a9335-cf15-4b30-ba2e-846e0daed72d-ca-trust-extracted\") pod \"d97a9335-cf15-4b30-ba2e-846e0daed72d\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.871811 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"d97a9335-cf15-4b30-ba2e-846e0daed72d\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.872025 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d97a9335-cf15-4b30-ba2e-846e0daed72d-registry-certificates\") pod \"d97a9335-cf15-4b30-ba2e-846e0daed72d\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.872122 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d97a9335-cf15-4b30-ba2e-846e0daed72d-installation-pull-secrets\") pod \"d97a9335-cf15-4b30-ba2e-846e0daed72d\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.872166 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpr69\" (UniqueName: \"kubernetes.io/projected/d97a9335-cf15-4b30-ba2e-846e0daed72d-kube-api-access-xpr69\") pod \"d97a9335-cf15-4b30-ba2e-846e0daed72d\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.872195 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d97a9335-cf15-4b30-ba2e-846e0daed72d-trusted-ca\") pod \"d97a9335-cf15-4b30-ba2e-846e0daed72d\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.872221 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d97a9335-cf15-4b30-ba2e-846e0daed72d-registry-tls\") pod \"d97a9335-cf15-4b30-ba2e-846e0daed72d\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.872244 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d97a9335-cf15-4b30-ba2e-846e0daed72d-bound-sa-token\") pod \"d97a9335-cf15-4b30-ba2e-846e0daed72d\" (UID: \"d97a9335-cf15-4b30-ba2e-846e0daed72d\") " Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.873086 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d97a9335-cf15-4b30-ba2e-846e0daed72d-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "d97a9335-cf15-4b30-ba2e-846e0daed72d" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.873410 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d97a9335-cf15-4b30-ba2e-846e0daed72d-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "d97a9335-cf15-4b30-ba2e-846e0daed72d" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.877524 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d97a9335-cf15-4b30-ba2e-846e0daed72d-kube-api-access-xpr69" (OuterVolumeSpecName: "kube-api-access-xpr69") pod "d97a9335-cf15-4b30-ba2e-846e0daed72d" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d"). InnerVolumeSpecName "kube-api-access-xpr69". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.881005 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d97a9335-cf15-4b30-ba2e-846e0daed72d-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "d97a9335-cf15-4b30-ba2e-846e0daed72d" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.881150 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d97a9335-cf15-4b30-ba2e-846e0daed72d-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "d97a9335-cf15-4b30-ba2e-846e0daed72d" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.882180 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d97a9335-cf15-4b30-ba2e-846e0daed72d-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "d97a9335-cf15-4b30-ba2e-846e0daed72d" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.887480 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d97a9335-cf15-4b30-ba2e-846e0daed72d-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "d97a9335-cf15-4b30-ba2e-846e0daed72d" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.911148 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "d97a9335-cf15-4b30-ba2e-846e0daed72d" (UID: "d97a9335-cf15-4b30-ba2e-846e0daed72d"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.973198 4904 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d97a9335-cf15-4b30-ba2e-846e0daed72d-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.973237 4904 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d97a9335-cf15-4b30-ba2e-846e0daed72d-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.973246 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpr69\" (UniqueName: \"kubernetes.io/projected/d97a9335-cf15-4b30-ba2e-846e0daed72d-kube-api-access-xpr69\") on node \"crc\" DevicePath \"\"" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.973255 4904 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d97a9335-cf15-4b30-ba2e-846e0daed72d-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.973264 4904 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d97a9335-cf15-4b30-ba2e-846e0daed72d-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.973271 4904 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d97a9335-cf15-4b30-ba2e-846e0daed72d-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 14 11:17:00 crc kubenswrapper[4904]: I0214 11:17:00.973281 4904 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d97a9335-cf15-4b30-ba2e-846e0daed72d-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 14 11:17:01 crc kubenswrapper[4904]: I0214 11:17:01.080428 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zbh4h"] Feb 14 11:17:01 crc kubenswrapper[4904]: I0214 11:17:01.085893 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zbh4h"] Feb 14 11:17:01 crc kubenswrapper[4904]: I0214 11:17:01.842216 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d97a9335-cf15-4b30-ba2e-846e0daed72d" path="/var/lib/kubelet/pods/d97a9335-cf15-4b30-ba2e-846e0daed72d/volumes" Feb 14 11:17:16 crc kubenswrapper[4904]: I0214 11:17:16.382818 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:17:16 crc kubenswrapper[4904]: I0214 11:17:16.383273 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:17:16 crc kubenswrapper[4904]: I0214 11:17:16.383317 4904 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:17:16 crc kubenswrapper[4904]: I0214 11:17:16.383817 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"903ea27dacf6775c4c9f46b3a1c3654b638e3bdce73cd947b4f81a7bea2eca49"} pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 14 11:17:16 crc kubenswrapper[4904]: I0214 11:17:16.383890 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" containerID="cri-o://903ea27dacf6775c4c9f46b3a1c3654b638e3bdce73cd947b4f81a7bea2eca49" gracePeriod=600 Feb 14 11:17:16 crc kubenswrapper[4904]: I0214 11:17:16.831530 4904 generic.go:334] "Generic (PLEG): container finished" podID="7790d276-2506-4641-8c22-67e6441417f0" containerID="903ea27dacf6775c4c9f46b3a1c3654b638e3bdce73cd947b4f81a7bea2eca49" exitCode=0 Feb 14 11:17:16 crc kubenswrapper[4904]: I0214 11:17:16.831606 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerDied","Data":"903ea27dacf6775c4c9f46b3a1c3654b638e3bdce73cd947b4f81a7bea2eca49"} Feb 14 11:17:16 crc kubenswrapper[4904]: I0214 11:17:16.831909 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"7bee02c84077b982480197f3573826a9bf5c596ca426336719e5c06a6c3588f2"} Feb 14 11:17:16 crc kubenswrapper[4904]: I0214 11:17:16.831936 4904 scope.go:117] "RemoveContainer" containerID="29fd2a980f09baa770f02419f58a3d39f3a74624b34cff612a2343db28b3c228" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.654023 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-zwf2w"] Feb 14 11:19:08 crc kubenswrapper[4904]: E0214 11:19:08.654810 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d97a9335-cf15-4b30-ba2e-846e0daed72d" containerName="registry" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.654824 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="d97a9335-cf15-4b30-ba2e-846e0daed72d" containerName="registry" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.654983 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="d97a9335-cf15-4b30-ba2e-846e0daed72d" containerName="registry" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.655421 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-zwf2w" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.657365 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-8qkmb"] Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.658013 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-8qkmb" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.666525 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.666662 4904 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-55fdc" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.667097 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.667447 4904 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-f5789" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.679791 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-zwf2w"] Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.710680 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-8qkmb"] Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.712449 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-rr2vn"] Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.713055 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-rr2vn" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.724059 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-rr2vn"] Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.727507 4904 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-zsbcc" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.826368 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5wk8\" (UniqueName: \"kubernetes.io/projected/25466585-eaf0-4056-9bd2-766da897ddf6-kube-api-access-l5wk8\") pod \"cert-manager-webhook-687f57d79b-rr2vn\" (UID: \"25466585-eaf0-4056-9bd2-766da897ddf6\") " pod="cert-manager/cert-manager-webhook-687f57d79b-rr2vn" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.826453 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ggwd\" (UniqueName: \"kubernetes.io/projected/34109571-3f63-4e10-b5ed-ed86c6713f24-kube-api-access-8ggwd\") pod \"cert-manager-858654f9db-8qkmb\" (UID: \"34109571-3f63-4e10-b5ed-ed86c6713f24\") " pod="cert-manager/cert-manager-858654f9db-8qkmb" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.826474 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9czh\" (UniqueName: \"kubernetes.io/projected/b643d341-7020-4a44-9c9a-bd60f9968f90-kube-api-access-t9czh\") pod \"cert-manager-cainjector-cf98fcc89-zwf2w\" (UID: \"b643d341-7020-4a44-9c9a-bd60f9968f90\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-zwf2w" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.929052 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ggwd\" (UniqueName: \"kubernetes.io/projected/34109571-3f63-4e10-b5ed-ed86c6713f24-kube-api-access-8ggwd\") pod \"cert-manager-858654f9db-8qkmb\" (UID: \"34109571-3f63-4e10-b5ed-ed86c6713f24\") " pod="cert-manager/cert-manager-858654f9db-8qkmb" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.929409 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9czh\" (UniqueName: \"kubernetes.io/projected/b643d341-7020-4a44-9c9a-bd60f9968f90-kube-api-access-t9czh\") pod \"cert-manager-cainjector-cf98fcc89-zwf2w\" (UID: \"b643d341-7020-4a44-9c9a-bd60f9968f90\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-zwf2w" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.929536 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5wk8\" (UniqueName: \"kubernetes.io/projected/25466585-eaf0-4056-9bd2-766da897ddf6-kube-api-access-l5wk8\") pod \"cert-manager-webhook-687f57d79b-rr2vn\" (UID: \"25466585-eaf0-4056-9bd2-766da897ddf6\") " pod="cert-manager/cert-manager-webhook-687f57d79b-rr2vn" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.948735 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5wk8\" (UniqueName: \"kubernetes.io/projected/25466585-eaf0-4056-9bd2-766da897ddf6-kube-api-access-l5wk8\") pod \"cert-manager-webhook-687f57d79b-rr2vn\" (UID: \"25466585-eaf0-4056-9bd2-766da897ddf6\") " pod="cert-manager/cert-manager-webhook-687f57d79b-rr2vn" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.949459 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ggwd\" (UniqueName: \"kubernetes.io/projected/34109571-3f63-4e10-b5ed-ed86c6713f24-kube-api-access-8ggwd\") pod \"cert-manager-858654f9db-8qkmb\" (UID: \"34109571-3f63-4e10-b5ed-ed86c6713f24\") " pod="cert-manager/cert-manager-858654f9db-8qkmb" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.950894 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9czh\" (UniqueName: \"kubernetes.io/projected/b643d341-7020-4a44-9c9a-bd60f9968f90-kube-api-access-t9czh\") pod \"cert-manager-cainjector-cf98fcc89-zwf2w\" (UID: \"b643d341-7020-4a44-9c9a-bd60f9968f90\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-zwf2w" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.976947 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-zwf2w" Feb 14 11:19:08 crc kubenswrapper[4904]: I0214 11:19:08.996722 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-8qkmb" Feb 14 11:19:09 crc kubenswrapper[4904]: I0214 11:19:09.025046 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-rr2vn" Feb 14 11:19:09 crc kubenswrapper[4904]: I0214 11:19:09.225760 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-zwf2w"] Feb 14 11:19:09 crc kubenswrapper[4904]: I0214 11:19:09.234349 4904 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 14 11:19:09 crc kubenswrapper[4904]: I0214 11:19:09.410773 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-zwf2w" event={"ID":"b643d341-7020-4a44-9c9a-bd60f9968f90","Type":"ContainerStarted","Data":"660343a2e7a5a14e862375b62e5df8a431d8c8dde23cb726c2d25a7cce2e3fdb"} Feb 14 11:19:09 crc kubenswrapper[4904]: I0214 11:19:09.468907 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-8qkmb"] Feb 14 11:19:09 crc kubenswrapper[4904]: W0214 11:19:09.472899 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34109571_3f63_4e10_b5ed_ed86c6713f24.slice/crio-b59d62be9b94537b6cb724695e9c915a89a87d3fa81d90aad1d8be6ed22296e1 WatchSource:0}: Error finding container b59d62be9b94537b6cb724695e9c915a89a87d3fa81d90aad1d8be6ed22296e1: Status 404 returned error can't find the container with id b59d62be9b94537b6cb724695e9c915a89a87d3fa81d90aad1d8be6ed22296e1 Feb 14 11:19:09 crc kubenswrapper[4904]: I0214 11:19:09.515270 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-rr2vn"] Feb 14 11:19:09 crc kubenswrapper[4904]: W0214 11:19:09.520737 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25466585_eaf0_4056_9bd2_766da897ddf6.slice/crio-6d60fa090f78affb52dba4e247c865e1167da7161e714dd8b95c54817730fbb6 WatchSource:0}: Error finding container 6d60fa090f78affb52dba4e247c865e1167da7161e714dd8b95c54817730fbb6: Status 404 returned error can't find the container with id 6d60fa090f78affb52dba4e247c865e1167da7161e714dd8b95c54817730fbb6 Feb 14 11:19:10 crc kubenswrapper[4904]: I0214 11:19:10.085452 4904 scope.go:117] "RemoveContainer" containerID="f9c8683559f28895e271f2eefc21b2d9b9350ccea0cb3b9f4b127458bf1cae33" Feb 14 11:19:10 crc kubenswrapper[4904]: I0214 11:19:10.420202 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-rr2vn" event={"ID":"25466585-eaf0-4056-9bd2-766da897ddf6","Type":"ContainerStarted","Data":"6d60fa090f78affb52dba4e247c865e1167da7161e714dd8b95c54817730fbb6"} Feb 14 11:19:10 crc kubenswrapper[4904]: I0214 11:19:10.422328 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-8qkmb" event={"ID":"34109571-3f63-4e10-b5ed-ed86c6713f24","Type":"ContainerStarted","Data":"b59d62be9b94537b6cb724695e9c915a89a87d3fa81d90aad1d8be6ed22296e1"} Feb 14 11:19:11 crc kubenswrapper[4904]: I0214 11:19:11.429144 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-zwf2w" event={"ID":"b643d341-7020-4a44-9c9a-bd60f9968f90","Type":"ContainerStarted","Data":"534f17faaa49ff5c842dc8d61ade6e6f5034a96a01ca7cf076038e5c328514f6"} Feb 14 11:19:11 crc kubenswrapper[4904]: I0214 11:19:11.450080 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-zwf2w" podStartSLOduration=1.509211086 podStartE2EDuration="3.450055888s" podCreationTimestamp="2026-02-14 11:19:08 +0000 UTC" firstStartedPulling="2026-02-14 11:19:09.234175814 +0000 UTC m=+540.046940475" lastFinishedPulling="2026-02-14 11:19:11.175020616 +0000 UTC m=+541.987785277" observedRunningTime="2026-02-14 11:19:11.445019503 +0000 UTC m=+542.257784164" watchObservedRunningTime="2026-02-14 11:19:11.450055888 +0000 UTC m=+542.262820559" Feb 14 11:19:13 crc kubenswrapper[4904]: I0214 11:19:13.439991 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-rr2vn" event={"ID":"25466585-eaf0-4056-9bd2-766da897ddf6","Type":"ContainerStarted","Data":"01d8387466eb1461d05cc6b8bf65f5b2d9a70be1e0eb128e29ff08d85d94d6a9"} Feb 14 11:19:13 crc kubenswrapper[4904]: I0214 11:19:13.440500 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-rr2vn" Feb 14 11:19:13 crc kubenswrapper[4904]: I0214 11:19:13.442540 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-8qkmb" event={"ID":"34109571-3f63-4e10-b5ed-ed86c6713f24","Type":"ContainerStarted","Data":"4fccc41c4c9892cc2ef8e894d926d19519d5e0c039a4568da104d3a10a18190c"} Feb 14 11:19:13 crc kubenswrapper[4904]: I0214 11:19:13.461896 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-rr2vn" podStartSLOduration=2.225461991 podStartE2EDuration="5.461873041s" podCreationTimestamp="2026-02-14 11:19:08 +0000 UTC" firstStartedPulling="2026-02-14 11:19:09.523467756 +0000 UTC m=+540.336232417" lastFinishedPulling="2026-02-14 11:19:12.759878806 +0000 UTC m=+543.572643467" observedRunningTime="2026-02-14 11:19:13.453755508 +0000 UTC m=+544.266520169" watchObservedRunningTime="2026-02-14 11:19:13.461873041 +0000 UTC m=+544.274637702" Feb 14 11:19:13 crc kubenswrapper[4904]: I0214 11:19:13.470248 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-8qkmb" podStartSLOduration=2.135262687 podStartE2EDuration="5.470217121s" podCreationTimestamp="2026-02-14 11:19:08 +0000 UTC" firstStartedPulling="2026-02-14 11:19:09.475351933 +0000 UTC m=+540.288116604" lastFinishedPulling="2026-02-14 11:19:12.810306337 +0000 UTC m=+543.623071038" observedRunningTime="2026-02-14 11:19:13.466866104 +0000 UTC m=+544.279630775" watchObservedRunningTime="2026-02-14 11:19:13.470217121 +0000 UTC m=+544.282981782" Feb 14 11:19:16 crc kubenswrapper[4904]: I0214 11:19:16.382722 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:19:16 crc kubenswrapper[4904]: I0214 11:19:16.382793 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.330816 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2gwlw"] Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.331749 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovn-controller" containerID="cri-o://b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a" gracePeriod=30 Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.331868 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72" gracePeriod=30 Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.331918 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="northd" containerID="cri-o://e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175" gracePeriod=30 Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.331930 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="sbdb" containerID="cri-o://0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db" gracePeriod=30 Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.331995 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="kube-rbac-proxy-node" containerID="cri-o://d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4" gracePeriod=30 Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.332018 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovn-acl-logging" containerID="cri-o://a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39" gracePeriod=30 Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.334293 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="nbdb" containerID="cri-o://2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27" gracePeriod=30 Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.365220 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovnkube-controller" containerID="cri-o://8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6" gracePeriod=30 Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.473362 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jz5f5_0b01cc02-2e62-46e1-b07d-b3c0bff2b050/kube-multus/2.log" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.474016 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jz5f5_0b01cc02-2e62-46e1-b07d-b3c0bff2b050/kube-multus/1.log" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.474054 4904 generic.go:334] "Generic (PLEG): container finished" podID="0b01cc02-2e62-46e1-b07d-b3c0bff2b050" containerID="a746e87d6bbe5102b3055a0f666d693b83f8f709786e5b8ab22d2cbaf9a6b7ee" exitCode=2 Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.474108 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jz5f5" event={"ID":"0b01cc02-2e62-46e1-b07d-b3c0bff2b050","Type":"ContainerDied","Data":"a746e87d6bbe5102b3055a0f666d693b83f8f709786e5b8ab22d2cbaf9a6b7ee"} Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.474139 4904 scope.go:117] "RemoveContainer" containerID="58ce2d3b3613f2f1c1ec36d61edb9bb614e5c1cdf37132359c08ef1f142d8bb0" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.475571 4904 scope.go:117] "RemoveContainer" containerID="a746e87d6bbe5102b3055a0f666d693b83f8f709786e5b8ab22d2cbaf9a6b7ee" Feb 14 11:19:18 crc kubenswrapper[4904]: E0214 11:19:18.476652 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-jz5f5_openshift-multus(0b01cc02-2e62-46e1-b07d-b3c0bff2b050)\"" pod="openshift-multus/multus-jz5f5" podUID="0b01cc02-2e62-46e1-b07d-b3c0bff2b050" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.478180 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovnkube-controller/3.log" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.504382 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovn-acl-logging/0.log" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.507705 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovn-controller/0.log" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.509000 4904 generic.go:334] "Generic (PLEG): container finished" podID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerID="047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72" exitCode=0 Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.509025 4904 generic.go:334] "Generic (PLEG): container finished" podID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerID="d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4" exitCode=0 Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.509031 4904 generic.go:334] "Generic (PLEG): container finished" podID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerID="a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39" exitCode=143 Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.509038 4904 generic.go:334] "Generic (PLEG): container finished" podID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerID="b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a" exitCode=143 Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.509059 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerDied","Data":"047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72"} Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.509084 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerDied","Data":"d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4"} Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.509095 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerDied","Data":"a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39"} Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.509105 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerDied","Data":"b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a"} Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.665732 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovnkube-controller/3.log" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.667942 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovn-acl-logging/0.log" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.668480 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovn-controller/0.log" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.668896 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.719486 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8bqx6"] Feb 14 11:19:18 crc kubenswrapper[4904]: E0214 11:19:18.719709 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="kubecfg-setup" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.719720 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="kubecfg-setup" Feb 14 11:19:18 crc kubenswrapper[4904]: E0214 11:19:18.719731 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="sbdb" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.719737 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="sbdb" Feb 14 11:19:18 crc kubenswrapper[4904]: E0214 11:19:18.719745 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovnkube-controller" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.719752 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovnkube-controller" Feb 14 11:19:18 crc kubenswrapper[4904]: E0214 11:19:18.719759 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovn-controller" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.719765 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovn-controller" Feb 14 11:19:18 crc kubenswrapper[4904]: E0214 11:19:18.719774 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovn-acl-logging" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.719780 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovn-acl-logging" Feb 14 11:19:18 crc kubenswrapper[4904]: E0214 11:19:18.719786 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="northd" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.719793 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="northd" Feb 14 11:19:18 crc kubenswrapper[4904]: E0214 11:19:18.719806 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="kube-rbac-proxy-ovn-metrics" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.719813 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="kube-rbac-proxy-ovn-metrics" Feb 14 11:19:18 crc kubenswrapper[4904]: E0214 11:19:18.719823 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovnkube-controller" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.719828 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovnkube-controller" Feb 14 11:19:18 crc kubenswrapper[4904]: E0214 11:19:18.720034 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovnkube-controller" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.720040 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovnkube-controller" Feb 14 11:19:18 crc kubenswrapper[4904]: E0214 11:19:18.720045 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="kube-rbac-proxy-node" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.720051 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="kube-rbac-proxy-node" Feb 14 11:19:18 crc kubenswrapper[4904]: E0214 11:19:18.720059 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="nbdb" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.720065 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="nbdb" Feb 14 11:19:18 crc kubenswrapper[4904]: E0214 11:19:18.720075 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovnkube-controller" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.720080 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovnkube-controller" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.720164 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="northd" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.720173 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="sbdb" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.720182 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovnkube-controller" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.720189 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="nbdb" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.720197 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovn-acl-logging" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.720207 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovnkube-controller" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.720214 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="kube-rbac-proxy-node" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.720221 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="kube-rbac-proxy-ovn-metrics" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.720227 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovn-controller" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.720235 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovnkube-controller" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.720242 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovnkube-controller" Feb 14 11:19:18 crc kubenswrapper[4904]: E0214 11:19:18.720336 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovnkube-controller" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.720343 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovnkube-controller" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.720429 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerName="ovnkube-controller" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.721881 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.774667 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-cni-netd\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.774722 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-run-ovn\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.774762 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1ca5acc0-8b15-4052-883f-1987b7ca0067-ovnkube-config\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.774795 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rghsz\" (UniqueName: \"kubernetes.io/projected/1ca5acc0-8b15-4052-883f-1987b7ca0067-kube-api-access-rghsz\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.774816 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.774859 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-node-log\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.774891 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-node-log" (OuterVolumeSpecName: "node-log") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.774895 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-systemd-units\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.774957 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-run-openvswitch\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.774893 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.774980 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.774950 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775000 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-run-systemd\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775051 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-cni-bin\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775073 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-log-socket\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775094 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-slash\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775107 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775116 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-run-netns\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775118 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-log-socket" (OuterVolumeSpecName: "log-socket") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775145 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-slash" (OuterVolumeSpecName: "host-slash") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775147 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1ca5acc0-8b15-4052-883f-1987b7ca0067-ovn-node-metrics-cert\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775188 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-kubelet\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775210 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-var-lib-cni-networks-ovn-kubernetes\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775231 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-etc-openvswitch\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775252 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-var-lib-openvswitch\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775280 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1ca5acc0-8b15-4052-883f-1987b7ca0067-env-overrides\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775306 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1ca5acc0-8b15-4052-883f-1987b7ca0067-ovnkube-script-lib\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775334 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-run-ovn-kubernetes\") pod \"1ca5acc0-8b15-4052-883f-1987b7ca0067\" (UID: \"1ca5acc0-8b15-4052-883f-1987b7ca0067\") " Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775282 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775299 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775315 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775333 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775347 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775443 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ca5acc0-8b15-4052-883f-1987b7ca0067-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775493 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775613 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ca5acc0-8b15-4052-883f-1987b7ca0067-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775797 4904 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-node-log\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775818 4904 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775851 4904 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775864 4904 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775876 4904 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-log-socket\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775890 4904 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-slash\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775901 4904 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775912 4904 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775927 4904 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775935 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ca5acc0-8b15-4052-883f-1987b7ca0067-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775942 4904 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775975 4904 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775986 4904 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1ca5acc0-8b15-4052-883f-1987b7ca0067-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.775999 4904 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.776011 4904 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.776022 4904 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.776032 4904 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1ca5acc0-8b15-4052-883f-1987b7ca0067-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.780181 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca5acc0-8b15-4052-883f-1987b7ca0067-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.780720 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ca5acc0-8b15-4052-883f-1987b7ca0067-kube-api-access-rghsz" (OuterVolumeSpecName: "kube-api-access-rghsz") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "kube-api-access-rghsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.787301 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "1ca5acc0-8b15-4052-883f-1987b7ca0067" (UID: "1ca5acc0-8b15-4052-883f-1987b7ca0067"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876533 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-run-ovn-kubernetes\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876609 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-ovn-node-metrics-cert\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876630 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-run-systemd\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876646 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-env-overrides\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876666 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-run-ovn\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876682 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-kubelet\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876696 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-cni-bin\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876713 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-cni-netd\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876734 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-run-netns\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876751 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-log-socket\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876764 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-run-openvswitch\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876788 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-systemd-units\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876814 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-var-lib-openvswitch\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876828 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-ovnkube-config\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876858 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kthv6\" (UniqueName: \"kubernetes.io/projected/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-kube-api-access-kthv6\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876878 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-slash\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876900 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-ovnkube-script-lib\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876925 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876950 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-etc-openvswitch\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876964 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-node-log\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.876991 4904 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1ca5acc0-8b15-4052-883f-1987b7ca0067-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.877001 4904 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1ca5acc0-8b15-4052-883f-1987b7ca0067-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.877009 4904 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1ca5acc0-8b15-4052-883f-1987b7ca0067-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.877018 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rghsz\" (UniqueName: \"kubernetes.io/projected/1ca5acc0-8b15-4052-883f-1987b7ca0067-kube-api-access-rghsz\") on node \"crc\" DevicePath \"\"" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.977923 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-run-ovn-kubernetes\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978261 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-run-ovn-kubernetes\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978263 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-ovn-node-metrics-cert\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978348 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-run-systemd\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978368 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-env-overrides\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978387 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-run-ovn\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978416 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-kubelet\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978439 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-cni-bin\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978461 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-cni-netd\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978496 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-run-netns\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978533 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-log-socket\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978555 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-run-openvswitch\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978585 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-systemd-units\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978618 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kthv6\" (UniqueName: \"kubernetes.io/projected/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-kube-api-access-kthv6\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978636 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-var-lib-openvswitch\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978651 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-ovnkube-config\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978677 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-slash\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978721 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-ovnkube-script-lib\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978761 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978785 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-etc-openvswitch\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978808 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-node-log\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978877 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-node-log\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978900 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-run-systemd\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.978985 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-systemd-units\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.979078 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-cni-netd\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.979094 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-run-ovn\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.979117 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-kubelet\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.979137 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-cni-bin\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.979160 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-var-lib-openvswitch\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.979342 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-env-overrides\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.979450 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-run-netns\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.979457 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-log-socket\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.979592 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-run-openvswitch\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.979673 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-etc-openvswitch\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.979751 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.979786 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-ovnkube-script-lib\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.979887 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-host-slash\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.980344 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-ovnkube-config\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.983456 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-ovn-node-metrics-cert\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:18 crc kubenswrapper[4904]: I0214 11:19:18.993647 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kthv6\" (UniqueName: \"kubernetes.io/projected/8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb-kube-api-access-kthv6\") pod \"ovnkube-node-8bqx6\" (UID: \"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.027389 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-rr2vn" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.036028 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.516344 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jz5f5_0b01cc02-2e62-46e1-b07d-b3c0bff2b050/kube-multus/2.log" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.517818 4904 generic.go:334] "Generic (PLEG): container finished" podID="8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb" containerID="193cce067842fcdba8effb9439fd8547a9bfbd34e145191cecb089ebd7a4275f" exitCode=0 Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.517912 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" event={"ID":"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb","Type":"ContainerDied","Data":"193cce067842fcdba8effb9439fd8547a9bfbd34e145191cecb089ebd7a4275f"} Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.517969 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" event={"ID":"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb","Type":"ContainerStarted","Data":"531a80694f6c5e180e26724ac7e24da204ebcb8e71a078afadca2a59732d2414"} Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.521955 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovnkube-controller/3.log" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.527033 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovn-acl-logging/0.log" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.527512 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gwlw_1ca5acc0-8b15-4052-883f-1987b7ca0067/ovn-controller/0.log" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.527856 4904 generic.go:334] "Generic (PLEG): container finished" podID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerID="8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6" exitCode=0 Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.528050 4904 generic.go:334] "Generic (PLEG): container finished" podID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerID="0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db" exitCode=0 Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.528116 4904 generic.go:334] "Generic (PLEG): container finished" podID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerID="2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27" exitCode=0 Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.528194 4904 generic.go:334] "Generic (PLEG): container finished" podID="1ca5acc0-8b15-4052-883f-1987b7ca0067" containerID="e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175" exitCode=0 Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.528086 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.528020 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerDied","Data":"8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6"} Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.528871 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerDied","Data":"0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db"} Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.528893 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerDied","Data":"2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27"} Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.528906 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerDied","Data":"e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175"} Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.528917 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gwlw" event={"ID":"1ca5acc0-8b15-4052-883f-1987b7ca0067","Type":"ContainerDied","Data":"410843315a7fa3be626cd1a1f58bd0f7ad7bfa79e8cf2e69e8a8103c5e86f49a"} Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.528934 4904 scope.go:117] "RemoveContainer" containerID="8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.616772 4904 scope.go:117] "RemoveContainer" containerID="9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.628296 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2gwlw"] Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.656492 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2gwlw"] Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.659613 4904 scope.go:117] "RemoveContainer" containerID="0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.678594 4904 scope.go:117] "RemoveContainer" containerID="2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.690026 4904 scope.go:117] "RemoveContainer" containerID="e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.720040 4904 scope.go:117] "RemoveContainer" containerID="047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.736998 4904 scope.go:117] "RemoveContainer" containerID="d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.750736 4904 scope.go:117] "RemoveContainer" containerID="a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.784119 4904 scope.go:117] "RemoveContainer" containerID="b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.804271 4904 scope.go:117] "RemoveContainer" containerID="11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.819354 4904 scope.go:117] "RemoveContainer" containerID="8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6" Feb 14 11:19:19 crc kubenswrapper[4904]: E0214 11:19:19.819767 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6\": container with ID starting with 8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6 not found: ID does not exist" containerID="8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.819808 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6"} err="failed to get container status \"8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6\": rpc error: code = NotFound desc = could not find container \"8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6\": container with ID starting with 8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.819856 4904 scope.go:117] "RemoveContainer" containerID="9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70" Feb 14 11:19:19 crc kubenswrapper[4904]: E0214 11:19:19.820184 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70\": container with ID starting with 9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70 not found: ID does not exist" containerID="9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.820205 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70"} err="failed to get container status \"9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70\": rpc error: code = NotFound desc = could not find container \"9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70\": container with ID starting with 9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.820221 4904 scope.go:117] "RemoveContainer" containerID="0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db" Feb 14 11:19:19 crc kubenswrapper[4904]: E0214 11:19:19.820442 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\": container with ID starting with 0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db not found: ID does not exist" containerID="0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.820459 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db"} err="failed to get container status \"0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\": rpc error: code = NotFound desc = could not find container \"0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\": container with ID starting with 0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.820470 4904 scope.go:117] "RemoveContainer" containerID="2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27" Feb 14 11:19:19 crc kubenswrapper[4904]: E0214 11:19:19.820796 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\": container with ID starting with 2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27 not found: ID does not exist" containerID="2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.820813 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27"} err="failed to get container status \"2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\": rpc error: code = NotFound desc = could not find container \"2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\": container with ID starting with 2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.820824 4904 scope.go:117] "RemoveContainer" containerID="e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175" Feb 14 11:19:19 crc kubenswrapper[4904]: E0214 11:19:19.821103 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\": container with ID starting with e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175 not found: ID does not exist" containerID="e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.821120 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175"} err="failed to get container status \"e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\": rpc error: code = NotFound desc = could not find container \"e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\": container with ID starting with e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.821130 4904 scope.go:117] "RemoveContainer" containerID="047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72" Feb 14 11:19:19 crc kubenswrapper[4904]: E0214 11:19:19.821303 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\": container with ID starting with 047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72 not found: ID does not exist" containerID="047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.821319 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72"} err="failed to get container status \"047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\": rpc error: code = NotFound desc = could not find container \"047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\": container with ID starting with 047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.821329 4904 scope.go:117] "RemoveContainer" containerID="d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4" Feb 14 11:19:19 crc kubenswrapper[4904]: E0214 11:19:19.821491 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\": container with ID starting with d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4 not found: ID does not exist" containerID="d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.821507 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4"} err="failed to get container status \"d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\": rpc error: code = NotFound desc = could not find container \"d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\": container with ID starting with d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.821520 4904 scope.go:117] "RemoveContainer" containerID="a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39" Feb 14 11:19:19 crc kubenswrapper[4904]: E0214 11:19:19.821664 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\": container with ID starting with a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39 not found: ID does not exist" containerID="a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.821726 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39"} err="failed to get container status \"a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\": rpc error: code = NotFound desc = could not find container \"a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\": container with ID starting with a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.821746 4904 scope.go:117] "RemoveContainer" containerID="b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a" Feb 14 11:19:19 crc kubenswrapper[4904]: E0214 11:19:19.821953 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\": container with ID starting with b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a not found: ID does not exist" containerID="b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.821974 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a"} err="failed to get container status \"b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\": rpc error: code = NotFound desc = could not find container \"b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\": container with ID starting with b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.821988 4904 scope.go:117] "RemoveContainer" containerID="11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e" Feb 14 11:19:19 crc kubenswrapper[4904]: E0214 11:19:19.822213 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\": container with ID starting with 11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e not found: ID does not exist" containerID="11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.822232 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e"} err="failed to get container status \"11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\": rpc error: code = NotFound desc = could not find container \"11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\": container with ID starting with 11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.822244 4904 scope.go:117] "RemoveContainer" containerID="8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.822481 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6"} err="failed to get container status \"8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6\": rpc error: code = NotFound desc = could not find container \"8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6\": container with ID starting with 8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.822497 4904 scope.go:117] "RemoveContainer" containerID="9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.822728 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70"} err="failed to get container status \"9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70\": rpc error: code = NotFound desc = could not find container \"9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70\": container with ID starting with 9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.822744 4904 scope.go:117] "RemoveContainer" containerID="0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.823162 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db"} err="failed to get container status \"0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\": rpc error: code = NotFound desc = could not find container \"0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\": container with ID starting with 0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.823180 4904 scope.go:117] "RemoveContainer" containerID="2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.823333 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27"} err="failed to get container status \"2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\": rpc error: code = NotFound desc = could not find container \"2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\": container with ID starting with 2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.823352 4904 scope.go:117] "RemoveContainer" containerID="e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.823518 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175"} err="failed to get container status \"e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\": rpc error: code = NotFound desc = could not find container \"e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\": container with ID starting with e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.823538 4904 scope.go:117] "RemoveContainer" containerID="047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.823740 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72"} err="failed to get container status \"047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\": rpc error: code = NotFound desc = could not find container \"047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\": container with ID starting with 047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.823755 4904 scope.go:117] "RemoveContainer" containerID="d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.824064 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4"} err="failed to get container status \"d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\": rpc error: code = NotFound desc = could not find container \"d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\": container with ID starting with d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.824083 4904 scope.go:117] "RemoveContainer" containerID="a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.824328 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39"} err="failed to get container status \"a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\": rpc error: code = NotFound desc = could not find container \"a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\": container with ID starting with a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.824347 4904 scope.go:117] "RemoveContainer" containerID="b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.824575 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a"} err="failed to get container status \"b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\": rpc error: code = NotFound desc = could not find container \"b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\": container with ID starting with b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.824611 4904 scope.go:117] "RemoveContainer" containerID="11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.824854 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e"} err="failed to get container status \"11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\": rpc error: code = NotFound desc = could not find container \"11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\": container with ID starting with 11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.824878 4904 scope.go:117] "RemoveContainer" containerID="8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.825126 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6"} err="failed to get container status \"8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6\": rpc error: code = NotFound desc = could not find container \"8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6\": container with ID starting with 8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.825150 4904 scope.go:117] "RemoveContainer" containerID="9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.825374 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70"} err="failed to get container status \"9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70\": rpc error: code = NotFound desc = could not find container \"9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70\": container with ID starting with 9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.825393 4904 scope.go:117] "RemoveContainer" containerID="0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.825622 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db"} err="failed to get container status \"0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\": rpc error: code = NotFound desc = could not find container \"0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\": container with ID starting with 0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.825643 4904 scope.go:117] "RemoveContainer" containerID="2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.825894 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27"} err="failed to get container status \"2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\": rpc error: code = NotFound desc = could not find container \"2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\": container with ID starting with 2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.825911 4904 scope.go:117] "RemoveContainer" containerID="e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.826521 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175"} err="failed to get container status \"e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\": rpc error: code = NotFound desc = could not find container \"e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\": container with ID starting with e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.826552 4904 scope.go:117] "RemoveContainer" containerID="047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.826891 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72"} err="failed to get container status \"047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\": rpc error: code = NotFound desc = could not find container \"047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\": container with ID starting with 047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.826949 4904 scope.go:117] "RemoveContainer" containerID="d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.827235 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4"} err="failed to get container status \"d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\": rpc error: code = NotFound desc = could not find container \"d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\": container with ID starting with d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.827256 4904 scope.go:117] "RemoveContainer" containerID="a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.827944 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39"} err="failed to get container status \"a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\": rpc error: code = NotFound desc = could not find container \"a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\": container with ID starting with a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.827966 4904 scope.go:117] "RemoveContainer" containerID="b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.828258 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a"} err="failed to get container status \"b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\": rpc error: code = NotFound desc = could not find container \"b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\": container with ID starting with b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.828280 4904 scope.go:117] "RemoveContainer" containerID="11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.828527 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e"} err="failed to get container status \"11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\": rpc error: code = NotFound desc = could not find container \"11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\": container with ID starting with 11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.828545 4904 scope.go:117] "RemoveContainer" containerID="8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.828898 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6"} err="failed to get container status \"8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6\": rpc error: code = NotFound desc = could not find container \"8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6\": container with ID starting with 8999a7b87f434943e45e805420d0e1113a6dfe77d77e8cf8baaa493350883ff6 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.828917 4904 scope.go:117] "RemoveContainer" containerID="9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.829149 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70"} err="failed to get container status \"9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70\": rpc error: code = NotFound desc = could not find container \"9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70\": container with ID starting with 9654df688474459154eddfa9e26a0324c6d8fa2bb37a414ab8780f8341de9a70 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.829172 4904 scope.go:117] "RemoveContainer" containerID="0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.829452 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db"} err="failed to get container status \"0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\": rpc error: code = NotFound desc = could not find container \"0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db\": container with ID starting with 0ce672d694764659a8b0887f3ff2b31a844c6cede667f7780954b2b54e8d84db not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.829471 4904 scope.go:117] "RemoveContainer" containerID="2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.829649 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27"} err="failed to get container status \"2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\": rpc error: code = NotFound desc = could not find container \"2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27\": container with ID starting with 2d115179a8fa2bdd3d533e78001bc90dcd6332f51e73a282c788641eed201d27 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.829669 4904 scope.go:117] "RemoveContainer" containerID="e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.829938 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175"} err="failed to get container status \"e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\": rpc error: code = NotFound desc = could not find container \"e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175\": container with ID starting with e5fdfcca7b6f4c439288cff438d5a2603851e033cd4691a9b99614cb1decc175 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.830655 4904 scope.go:117] "RemoveContainer" containerID="047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.830925 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72"} err="failed to get container status \"047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\": rpc error: code = NotFound desc = could not find container \"047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72\": container with ID starting with 047e14641b269ae402fc7c1289a0e1f3145ebf849ca5b4f81f65b81f618e9b72 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.830944 4904 scope.go:117] "RemoveContainer" containerID="d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.831107 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4"} err="failed to get container status \"d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\": rpc error: code = NotFound desc = could not find container \"d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4\": container with ID starting with d64ae9b57af9eba798e5bd1ef4838a3dad0d3d52ac0ea2bbd13300b33a2e78f4 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.831124 4904 scope.go:117] "RemoveContainer" containerID="a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.831364 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39"} err="failed to get container status \"a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\": rpc error: code = NotFound desc = could not find container \"a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39\": container with ID starting with a8792a8d5c7c55f6a51a1b4af00eafc98ac9e112901d7fc4c8d8a7a44b021a39 not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.831381 4904 scope.go:117] "RemoveContainer" containerID="b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.831533 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a"} err="failed to get container status \"b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\": rpc error: code = NotFound desc = could not find container \"b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a\": container with ID starting with b17c84bd91bad6b837007a08548f8ef9bc79bf973a6ecaf87bb336055005b61a not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.831553 4904 scope.go:117] "RemoveContainer" containerID="11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.831715 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e"} err="failed to get container status \"11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\": rpc error: code = NotFound desc = could not find container \"11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e\": container with ID starting with 11bab7782ae819d3342cbceb206d69bccf2d8a15a29012746180abb25f37c66e not found: ID does not exist" Feb 14 11:19:19 crc kubenswrapper[4904]: I0214 11:19:19.850343 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ca5acc0-8b15-4052-883f-1987b7ca0067" path="/var/lib/kubelet/pods/1ca5acc0-8b15-4052-883f-1987b7ca0067/volumes" Feb 14 11:19:20 crc kubenswrapper[4904]: I0214 11:19:20.538053 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" event={"ID":"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb","Type":"ContainerStarted","Data":"f26f38d62505e88ef297fc62e5cde451b81c268304b99d532ca5d9d927de633e"} Feb 14 11:19:20 crc kubenswrapper[4904]: I0214 11:19:20.538310 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" event={"ID":"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb","Type":"ContainerStarted","Data":"c641207e1274a2e7739397ec5040a6b06cebceefe2747e47718200390c397945"} Feb 14 11:19:20 crc kubenswrapper[4904]: I0214 11:19:20.538323 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" event={"ID":"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb","Type":"ContainerStarted","Data":"1c6ab2a475fb46f0e13964951161a5b73680175374389bb5434a341158fcd6da"} Feb 14 11:19:20 crc kubenswrapper[4904]: I0214 11:19:20.538331 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" event={"ID":"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb","Type":"ContainerStarted","Data":"98c1796364a77a4f8aa4dda922f06e6ee55879a95020c154888e919f6bb619a3"} Feb 14 11:19:20 crc kubenswrapper[4904]: I0214 11:19:20.538339 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" event={"ID":"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb","Type":"ContainerStarted","Data":"e909401eef8875c202d85e6eebf3993f8977069cf6672d5fe08a2eac17632982"} Feb 14 11:19:20 crc kubenswrapper[4904]: I0214 11:19:20.538347 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" event={"ID":"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb","Type":"ContainerStarted","Data":"513b6af7393dafc8fdcb62ca48e1fbd182577050d4e5863dcff33eff04e5ab5f"} Feb 14 11:19:22 crc kubenswrapper[4904]: I0214 11:19:22.561997 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" event={"ID":"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb","Type":"ContainerStarted","Data":"66697a9c32ea9090f28bc81791273fff3389c2f9411c1dcc02a861bc255fd795"} Feb 14 11:19:25 crc kubenswrapper[4904]: I0214 11:19:25.586446 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" event={"ID":"8b16e2e7-2b8b-46cd-a7ab-b78b0acf20fb","Type":"ContainerStarted","Data":"85633f8c94107d2284a09bb692615d5f40bbb275584f654572d72d761528cab2"} Feb 14 11:19:25 crc kubenswrapper[4904]: I0214 11:19:25.587083 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:25 crc kubenswrapper[4904]: I0214 11:19:25.587196 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:25 crc kubenswrapper[4904]: I0214 11:19:25.587271 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:25 crc kubenswrapper[4904]: I0214 11:19:25.611082 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:25 crc kubenswrapper[4904]: I0214 11:19:25.612321 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:25 crc kubenswrapper[4904]: I0214 11:19:25.622869 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" podStartSLOduration=7.622851082 podStartE2EDuration="7.622851082s" podCreationTimestamp="2026-02-14 11:19:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:19:25.618006283 +0000 UTC m=+556.430770954" watchObservedRunningTime="2026-02-14 11:19:25.622851082 +0000 UTC m=+556.435615743" Feb 14 11:19:30 crc kubenswrapper[4904]: I0214 11:19:30.836239 4904 scope.go:117] "RemoveContainer" containerID="a746e87d6bbe5102b3055a0f666d693b83f8f709786e5b8ab22d2cbaf9a6b7ee" Feb 14 11:19:30 crc kubenswrapper[4904]: E0214 11:19:30.836900 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-jz5f5_openshift-multus(0b01cc02-2e62-46e1-b07d-b3c0bff2b050)\"" pod="openshift-multus/multus-jz5f5" podUID="0b01cc02-2e62-46e1-b07d-b3c0bff2b050" Feb 14 11:19:42 crc kubenswrapper[4904]: I0214 11:19:42.835821 4904 scope.go:117] "RemoveContainer" containerID="a746e87d6bbe5102b3055a0f666d693b83f8f709786e5b8ab22d2cbaf9a6b7ee" Feb 14 11:19:43 crc kubenswrapper[4904]: I0214 11:19:43.733658 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jz5f5_0b01cc02-2e62-46e1-b07d-b3c0bff2b050/kube-multus/2.log" Feb 14 11:19:43 crc kubenswrapper[4904]: I0214 11:19:43.734054 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jz5f5" event={"ID":"0b01cc02-2e62-46e1-b07d-b3c0bff2b050","Type":"ContainerStarted","Data":"3b5c42aa21ca0240d9997db64878fa9edcfd5e1cd7ccb8dda73eb2d64bdbb4b8"} Feb 14 11:19:46 crc kubenswrapper[4904]: I0214 11:19:46.382677 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:19:46 crc kubenswrapper[4904]: I0214 11:19:46.383010 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:19:49 crc kubenswrapper[4904]: I0214 11:19:49.056970 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8bqx6" Feb 14 11:19:59 crc kubenswrapper[4904]: I0214 11:19:59.136029 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn"] Feb 14 11:19:59 crc kubenswrapper[4904]: I0214 11:19:59.137899 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn" Feb 14 11:19:59 crc kubenswrapper[4904]: I0214 11:19:59.156697 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 14 11:19:59 crc kubenswrapper[4904]: I0214 11:19:59.163191 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn"] Feb 14 11:19:59 crc kubenswrapper[4904]: I0214 11:19:59.166086 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/becbe5d6-16f4-4f62-8c90-70be78f0c091-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn\" (UID: \"becbe5d6-16f4-4f62-8c90-70be78f0c091\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn" Feb 14 11:19:59 crc kubenswrapper[4904]: I0214 11:19:59.166161 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/becbe5d6-16f4-4f62-8c90-70be78f0c091-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn\" (UID: \"becbe5d6-16f4-4f62-8c90-70be78f0c091\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn" Feb 14 11:19:59 crc kubenswrapper[4904]: I0214 11:19:59.166204 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5mk7\" (UniqueName: \"kubernetes.io/projected/becbe5d6-16f4-4f62-8c90-70be78f0c091-kube-api-access-k5mk7\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn\" (UID: \"becbe5d6-16f4-4f62-8c90-70be78f0c091\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn" Feb 14 11:19:59 crc kubenswrapper[4904]: I0214 11:19:59.267014 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/becbe5d6-16f4-4f62-8c90-70be78f0c091-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn\" (UID: \"becbe5d6-16f4-4f62-8c90-70be78f0c091\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn" Feb 14 11:19:59 crc kubenswrapper[4904]: I0214 11:19:59.267092 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5mk7\" (UniqueName: \"kubernetes.io/projected/becbe5d6-16f4-4f62-8c90-70be78f0c091-kube-api-access-k5mk7\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn\" (UID: \"becbe5d6-16f4-4f62-8c90-70be78f0c091\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn" Feb 14 11:19:59 crc kubenswrapper[4904]: I0214 11:19:59.267145 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/becbe5d6-16f4-4f62-8c90-70be78f0c091-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn\" (UID: \"becbe5d6-16f4-4f62-8c90-70be78f0c091\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn" Feb 14 11:19:59 crc kubenswrapper[4904]: I0214 11:19:59.267701 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/becbe5d6-16f4-4f62-8c90-70be78f0c091-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn\" (UID: \"becbe5d6-16f4-4f62-8c90-70be78f0c091\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn" Feb 14 11:19:59 crc kubenswrapper[4904]: I0214 11:19:59.267700 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/becbe5d6-16f4-4f62-8c90-70be78f0c091-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn\" (UID: \"becbe5d6-16f4-4f62-8c90-70be78f0c091\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn" Feb 14 11:19:59 crc kubenswrapper[4904]: I0214 11:19:59.284106 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5mk7\" (UniqueName: \"kubernetes.io/projected/becbe5d6-16f4-4f62-8c90-70be78f0c091-kube-api-access-k5mk7\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn\" (UID: \"becbe5d6-16f4-4f62-8c90-70be78f0c091\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn" Feb 14 11:19:59 crc kubenswrapper[4904]: I0214 11:19:59.455173 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn" Feb 14 11:19:59 crc kubenswrapper[4904]: I0214 11:19:59.642642 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn"] Feb 14 11:19:59 crc kubenswrapper[4904]: W0214 11:19:59.653325 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbecbe5d6_16f4_4f62_8c90_70be78f0c091.slice/crio-bb6a78a64873850874b02ad4ce399c47147a65ab575ff04dfe8746f32a61fc65 WatchSource:0}: Error finding container bb6a78a64873850874b02ad4ce399c47147a65ab575ff04dfe8746f32a61fc65: Status 404 returned error can't find the container with id bb6a78a64873850874b02ad4ce399c47147a65ab575ff04dfe8746f32a61fc65 Feb 14 11:20:00 crc kubenswrapper[4904]: I0214 11:20:00.434544 4904 generic.go:334] "Generic (PLEG): container finished" podID="becbe5d6-16f4-4f62-8c90-70be78f0c091" containerID="482920c57a1f5cd6e1ddb26239f15445737bfe5cb38131251c2ade092764bac5" exitCode=0 Feb 14 11:20:00 crc kubenswrapper[4904]: I0214 11:20:00.434594 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn" event={"ID":"becbe5d6-16f4-4f62-8c90-70be78f0c091","Type":"ContainerDied","Data":"482920c57a1f5cd6e1ddb26239f15445737bfe5cb38131251c2ade092764bac5"} Feb 14 11:20:00 crc kubenswrapper[4904]: I0214 11:20:00.434624 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn" event={"ID":"becbe5d6-16f4-4f62-8c90-70be78f0c091","Type":"ContainerStarted","Data":"bb6a78a64873850874b02ad4ce399c47147a65ab575ff04dfe8746f32a61fc65"} Feb 14 11:20:02 crc kubenswrapper[4904]: I0214 11:20:02.455887 4904 generic.go:334] "Generic (PLEG): container finished" podID="becbe5d6-16f4-4f62-8c90-70be78f0c091" containerID="9a095d46876cbc4228efb886ba73c687069f1aa68db0a97cc14e7b27e552091f" exitCode=0 Feb 14 11:20:02 crc kubenswrapper[4904]: I0214 11:20:02.456256 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn" event={"ID":"becbe5d6-16f4-4f62-8c90-70be78f0c091","Type":"ContainerDied","Data":"9a095d46876cbc4228efb886ba73c687069f1aa68db0a97cc14e7b27e552091f"} Feb 14 11:20:03 crc kubenswrapper[4904]: I0214 11:20:03.464152 4904 generic.go:334] "Generic (PLEG): container finished" podID="becbe5d6-16f4-4f62-8c90-70be78f0c091" containerID="53b31e94e887e6a3ae9615be6c6ed7744fd979dc689bbff4f164cb6f2acf4d49" exitCode=0 Feb 14 11:20:03 crc kubenswrapper[4904]: I0214 11:20:03.464285 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn" event={"ID":"becbe5d6-16f4-4f62-8c90-70be78f0c091","Type":"ContainerDied","Data":"53b31e94e887e6a3ae9615be6c6ed7744fd979dc689bbff4f164cb6f2acf4d49"} Feb 14 11:20:04 crc kubenswrapper[4904]: I0214 11:20:04.654244 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn" Feb 14 11:20:04 crc kubenswrapper[4904]: I0214 11:20:04.728231 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5mk7\" (UniqueName: \"kubernetes.io/projected/becbe5d6-16f4-4f62-8c90-70be78f0c091-kube-api-access-k5mk7\") pod \"becbe5d6-16f4-4f62-8c90-70be78f0c091\" (UID: \"becbe5d6-16f4-4f62-8c90-70be78f0c091\") " Feb 14 11:20:04 crc kubenswrapper[4904]: I0214 11:20:04.728300 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/becbe5d6-16f4-4f62-8c90-70be78f0c091-util\") pod \"becbe5d6-16f4-4f62-8c90-70be78f0c091\" (UID: \"becbe5d6-16f4-4f62-8c90-70be78f0c091\") " Feb 14 11:20:04 crc kubenswrapper[4904]: I0214 11:20:04.728323 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/becbe5d6-16f4-4f62-8c90-70be78f0c091-bundle\") pod \"becbe5d6-16f4-4f62-8c90-70be78f0c091\" (UID: \"becbe5d6-16f4-4f62-8c90-70be78f0c091\") " Feb 14 11:20:04 crc kubenswrapper[4904]: I0214 11:20:04.729049 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/becbe5d6-16f4-4f62-8c90-70be78f0c091-bundle" (OuterVolumeSpecName: "bundle") pod "becbe5d6-16f4-4f62-8c90-70be78f0c091" (UID: "becbe5d6-16f4-4f62-8c90-70be78f0c091"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:20:04 crc kubenswrapper[4904]: I0214 11:20:04.733121 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/becbe5d6-16f4-4f62-8c90-70be78f0c091-kube-api-access-k5mk7" (OuterVolumeSpecName: "kube-api-access-k5mk7") pod "becbe5d6-16f4-4f62-8c90-70be78f0c091" (UID: "becbe5d6-16f4-4f62-8c90-70be78f0c091"). InnerVolumeSpecName "kube-api-access-k5mk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:20:04 crc kubenswrapper[4904]: I0214 11:20:04.742234 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/becbe5d6-16f4-4f62-8c90-70be78f0c091-util" (OuterVolumeSpecName: "util") pod "becbe5d6-16f4-4f62-8c90-70be78f0c091" (UID: "becbe5d6-16f4-4f62-8c90-70be78f0c091"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:20:04 crc kubenswrapper[4904]: I0214 11:20:04.829976 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5mk7\" (UniqueName: \"kubernetes.io/projected/becbe5d6-16f4-4f62-8c90-70be78f0c091-kube-api-access-k5mk7\") on node \"crc\" DevicePath \"\"" Feb 14 11:20:04 crc kubenswrapper[4904]: I0214 11:20:04.830009 4904 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/becbe5d6-16f4-4f62-8c90-70be78f0c091-util\") on node \"crc\" DevicePath \"\"" Feb 14 11:20:04 crc kubenswrapper[4904]: I0214 11:20:04.830022 4904 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/becbe5d6-16f4-4f62-8c90-70be78f0c091-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:20:05 crc kubenswrapper[4904]: I0214 11:20:05.475561 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn" event={"ID":"becbe5d6-16f4-4f62-8c90-70be78f0c091","Type":"ContainerDied","Data":"bb6a78a64873850874b02ad4ce399c47147a65ab575ff04dfe8746f32a61fc65"} Feb 14 11:20:05 crc kubenswrapper[4904]: I0214 11:20:05.475883 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb6a78a64873850874b02ad4ce399c47147a65ab575ff04dfe8746f32a61fc65" Feb 14 11:20:05 crc kubenswrapper[4904]: I0214 11:20:05.475615 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn" Feb 14 11:20:06 crc kubenswrapper[4904]: I0214 11:20:06.938193 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-nxxlw"] Feb 14 11:20:06 crc kubenswrapper[4904]: E0214 11:20:06.938492 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="becbe5d6-16f4-4f62-8c90-70be78f0c091" containerName="pull" Feb 14 11:20:06 crc kubenswrapper[4904]: I0214 11:20:06.938514 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="becbe5d6-16f4-4f62-8c90-70be78f0c091" containerName="pull" Feb 14 11:20:06 crc kubenswrapper[4904]: E0214 11:20:06.938534 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="becbe5d6-16f4-4f62-8c90-70be78f0c091" containerName="extract" Feb 14 11:20:06 crc kubenswrapper[4904]: I0214 11:20:06.938545 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="becbe5d6-16f4-4f62-8c90-70be78f0c091" containerName="extract" Feb 14 11:20:06 crc kubenswrapper[4904]: E0214 11:20:06.938559 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="becbe5d6-16f4-4f62-8c90-70be78f0c091" containerName="util" Feb 14 11:20:06 crc kubenswrapper[4904]: I0214 11:20:06.938570 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="becbe5d6-16f4-4f62-8c90-70be78f0c091" containerName="util" Feb 14 11:20:06 crc kubenswrapper[4904]: I0214 11:20:06.938736 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="becbe5d6-16f4-4f62-8c90-70be78f0c091" containerName="extract" Feb 14 11:20:06 crc kubenswrapper[4904]: I0214 11:20:06.939363 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-nxxlw" Feb 14 11:20:06 crc kubenswrapper[4904]: I0214 11:20:06.941652 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-dcn2d" Feb 14 11:20:06 crc kubenswrapper[4904]: I0214 11:20:06.941768 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Feb 14 11:20:06 crc kubenswrapper[4904]: I0214 11:20:06.944749 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Feb 14 11:20:06 crc kubenswrapper[4904]: I0214 11:20:06.953335 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szg8g\" (UniqueName: \"kubernetes.io/projected/38f7c921-1f4f-4217-92db-94809df6750c-kube-api-access-szg8g\") pod \"nmstate-operator-694c9596b7-nxxlw\" (UID: \"38f7c921-1f4f-4217-92db-94809df6750c\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-nxxlw" Feb 14 11:20:06 crc kubenswrapper[4904]: I0214 11:20:06.961103 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-nxxlw"] Feb 14 11:20:07 crc kubenswrapper[4904]: I0214 11:20:07.054314 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szg8g\" (UniqueName: \"kubernetes.io/projected/38f7c921-1f4f-4217-92db-94809df6750c-kube-api-access-szg8g\") pod \"nmstate-operator-694c9596b7-nxxlw\" (UID: \"38f7c921-1f4f-4217-92db-94809df6750c\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-nxxlw" Feb 14 11:20:07 crc kubenswrapper[4904]: I0214 11:20:07.076065 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szg8g\" (UniqueName: \"kubernetes.io/projected/38f7c921-1f4f-4217-92db-94809df6750c-kube-api-access-szg8g\") pod \"nmstate-operator-694c9596b7-nxxlw\" (UID: \"38f7c921-1f4f-4217-92db-94809df6750c\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-nxxlw" Feb 14 11:20:07 crc kubenswrapper[4904]: I0214 11:20:07.253989 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-nxxlw" Feb 14 11:20:07 crc kubenswrapper[4904]: I0214 11:20:07.720210 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-nxxlw"] Feb 14 11:20:07 crc kubenswrapper[4904]: W0214 11:20:07.722875 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38f7c921_1f4f_4217_92db_94809df6750c.slice/crio-833ac16853524f39d8a080c60fb341bc12239a651d4a6242f2293b1bb7da1128 WatchSource:0}: Error finding container 833ac16853524f39d8a080c60fb341bc12239a651d4a6242f2293b1bb7da1128: Status 404 returned error can't find the container with id 833ac16853524f39d8a080c60fb341bc12239a651d4a6242f2293b1bb7da1128 Feb 14 11:20:08 crc kubenswrapper[4904]: I0214 11:20:08.489412 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-nxxlw" event={"ID":"38f7c921-1f4f-4217-92db-94809df6750c","Type":"ContainerStarted","Data":"833ac16853524f39d8a080c60fb341bc12239a651d4a6242f2293b1bb7da1128"} Feb 14 11:20:10 crc kubenswrapper[4904]: I0214 11:20:10.259183 4904 scope.go:117] "RemoveContainer" containerID="778442d1f9bdc3c507a90be7b2083d3b93f728b5a9e12f99fef04a8a5601fc6a" Feb 14 11:20:10 crc kubenswrapper[4904]: I0214 11:20:10.280892 4904 scope.go:117] "RemoveContainer" containerID="e4ffd53226b042342190efa7d0897a5537d1ac98dca6af334e95d851e7c4e81f" Feb 14 11:20:10 crc kubenswrapper[4904]: I0214 11:20:10.501025 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-nxxlw" event={"ID":"38f7c921-1f4f-4217-92db-94809df6750c","Type":"ContainerStarted","Data":"6b39e571bc99717a563def6d3b807a3b10ae0afc3c582cf1bef0cb3e8ad125c7"} Feb 14 11:20:10 crc kubenswrapper[4904]: I0214 11:20:10.516561 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-694c9596b7-nxxlw" podStartSLOduration=2.786732986 podStartE2EDuration="4.516540521s" podCreationTimestamp="2026-02-14 11:20:06 +0000 UTC" firstStartedPulling="2026-02-14 11:20:07.726453453 +0000 UTC m=+598.539218114" lastFinishedPulling="2026-02-14 11:20:09.456260988 +0000 UTC m=+600.269025649" observedRunningTime="2026-02-14 11:20:10.513033019 +0000 UTC m=+601.325797680" watchObservedRunningTime="2026-02-14 11:20:10.516540521 +0000 UTC m=+601.329305182" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.427651 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-js9s2"] Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.428692 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-js9s2" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.434642 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-bsnv9" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.450908 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-js9s2"] Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.454365 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-tj7n4"] Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.455044 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tj7n4" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.457299 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.464023 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-g5z5r"] Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.464931 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-g5z5r" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.512136 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/62ca493b-5a80-450e-9e45-8f916f2c05e8-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-tj7n4\" (UID: \"62ca493b-5a80-450e-9e45-8f916f2c05e8\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tj7n4" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.512182 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzxc7\" (UniqueName: \"kubernetes.io/projected/62ca493b-5a80-450e-9e45-8f916f2c05e8-kube-api-access-gzxc7\") pod \"nmstate-webhook-866bcb46dc-tj7n4\" (UID: \"62ca493b-5a80-450e-9e45-8f916f2c05e8\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tj7n4" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.512357 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9zkg\" (UniqueName: \"kubernetes.io/projected/87d08820-f798-415f-a367-e1ec7bfb86d2-kube-api-access-w9zkg\") pod \"nmstate-metrics-58c85c668d-js9s2\" (UID: \"87d08820-f798-415f-a367-e1ec7bfb86d2\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-js9s2" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.512456 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62-nmstate-lock\") pod \"nmstate-handler-g5z5r\" (UID: \"967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62\") " pod="openshift-nmstate/nmstate-handler-g5z5r" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.512556 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62-dbus-socket\") pod \"nmstate-handler-g5z5r\" (UID: \"967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62\") " pod="openshift-nmstate/nmstate-handler-g5z5r" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.512595 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqz5h\" (UniqueName: \"kubernetes.io/projected/967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62-kube-api-access-mqz5h\") pod \"nmstate-handler-g5z5r\" (UID: \"967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62\") " pod="openshift-nmstate/nmstate-handler-g5z5r" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.512615 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62-ovs-socket\") pod \"nmstate-handler-g5z5r\" (UID: \"967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62\") " pod="openshift-nmstate/nmstate-handler-g5z5r" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.549109 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-tj7n4"] Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.602107 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-pc5mk"] Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.602813 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-pc5mk" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.604798 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.604805 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-8cxj9" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.611083 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.613630 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62-nmstate-lock\") pod \"nmstate-handler-g5z5r\" (UID: \"967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62\") " pod="openshift-nmstate/nmstate-handler-g5z5r" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.613679 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62-dbus-socket\") pod \"nmstate-handler-g5z5r\" (UID: \"967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62\") " pod="openshift-nmstate/nmstate-handler-g5z5r" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.613702 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqz5h\" (UniqueName: \"kubernetes.io/projected/967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62-kube-api-access-mqz5h\") pod \"nmstate-handler-g5z5r\" (UID: \"967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62\") " pod="openshift-nmstate/nmstate-handler-g5z5r" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.613720 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62-ovs-socket\") pod \"nmstate-handler-g5z5r\" (UID: \"967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62\") " pod="openshift-nmstate/nmstate-handler-g5z5r" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.613751 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/62ca493b-5a80-450e-9e45-8f916f2c05e8-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-tj7n4\" (UID: \"62ca493b-5a80-450e-9e45-8f916f2c05e8\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tj7n4" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.613767 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62-nmstate-lock\") pod \"nmstate-handler-g5z5r\" (UID: \"967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62\") " pod="openshift-nmstate/nmstate-handler-g5z5r" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.613792 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzxc7\" (UniqueName: \"kubernetes.io/projected/62ca493b-5a80-450e-9e45-8f916f2c05e8-kube-api-access-gzxc7\") pod \"nmstate-webhook-866bcb46dc-tj7n4\" (UID: \"62ca493b-5a80-450e-9e45-8f916f2c05e8\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tj7n4" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.613817 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62-ovs-socket\") pod \"nmstate-handler-g5z5r\" (UID: \"967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62\") " pod="openshift-nmstate/nmstate-handler-g5z5r" Feb 14 11:20:11 crc kubenswrapper[4904]: E0214 11:20:11.613909 4904 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Feb 14 11:20:11 crc kubenswrapper[4904]: E0214 11:20:11.613957 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62ca493b-5a80-450e-9e45-8f916f2c05e8-tls-key-pair podName:62ca493b-5a80-450e-9e45-8f916f2c05e8 nodeName:}" failed. No retries permitted until 2026-02-14 11:20:12.113941854 +0000 UTC m=+602.926706515 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/62ca493b-5a80-450e-9e45-8f916f2c05e8-tls-key-pair") pod "nmstate-webhook-866bcb46dc-tj7n4" (UID: "62ca493b-5a80-450e-9e45-8f916f2c05e8") : secret "openshift-nmstate-webhook" not found Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.614016 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9zkg\" (UniqueName: \"kubernetes.io/projected/87d08820-f798-415f-a367-e1ec7bfb86d2-kube-api-access-w9zkg\") pod \"nmstate-metrics-58c85c668d-js9s2\" (UID: \"87d08820-f798-415f-a367-e1ec7bfb86d2\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-js9s2" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.614409 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62-dbus-socket\") pod \"nmstate-handler-g5z5r\" (UID: \"967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62\") " pod="openshift-nmstate/nmstate-handler-g5z5r" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.627369 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-pc5mk"] Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.638205 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqz5h\" (UniqueName: \"kubernetes.io/projected/967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62-kube-api-access-mqz5h\") pod \"nmstate-handler-g5z5r\" (UID: \"967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62\") " pod="openshift-nmstate/nmstate-handler-g5z5r" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.641617 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzxc7\" (UniqueName: \"kubernetes.io/projected/62ca493b-5a80-450e-9e45-8f916f2c05e8-kube-api-access-gzxc7\") pod \"nmstate-webhook-866bcb46dc-tj7n4\" (UID: \"62ca493b-5a80-450e-9e45-8f916f2c05e8\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tj7n4" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.660509 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9zkg\" (UniqueName: \"kubernetes.io/projected/87d08820-f798-415f-a367-e1ec7bfb86d2-kube-api-access-w9zkg\") pod \"nmstate-metrics-58c85c668d-js9s2\" (UID: \"87d08820-f798-415f-a367-e1ec7bfb86d2\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-js9s2" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.715500 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-pc5mk\" (UID: \"3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-pc5mk" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.715579 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-pc5mk\" (UID: \"3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-pc5mk" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.715614 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fqtf\" (UniqueName: \"kubernetes.io/projected/3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a-kube-api-access-6fqtf\") pod \"nmstate-console-plugin-5c78fc5d65-pc5mk\" (UID: \"3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-pc5mk" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.744971 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-js9s2" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.796680 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-g5z5r" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.816797 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-pc5mk\" (UID: \"3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-pc5mk" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.817019 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-pc5mk\" (UID: \"3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-pc5mk" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.818298 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-pc5mk\" (UID: \"3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-pc5mk" Feb 14 11:20:11 crc kubenswrapper[4904]: E0214 11:20:11.816961 4904 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Feb 14 11:20:11 crc kubenswrapper[4904]: E0214 11:20:11.818893 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a-plugin-serving-cert podName:3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a nodeName:}" failed. No retries permitted until 2026-02-14 11:20:12.318877177 +0000 UTC m=+603.131641838 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a-plugin-serving-cert") pod "nmstate-console-plugin-5c78fc5d65-pc5mk" (UID: "3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a") : secret "plugin-serving-cert" not found Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.819044 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fqtf\" (UniqueName: \"kubernetes.io/projected/3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a-kube-api-access-6fqtf\") pod \"nmstate-console-plugin-5c78fc5d65-pc5mk\" (UID: \"3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-pc5mk" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.819331 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6456ccfc99-zhc2h"] Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.820059 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.849388 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fqtf\" (UniqueName: \"kubernetes.io/projected/3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a-kube-api-access-6fqtf\") pod \"nmstate-console-plugin-5c78fc5d65-pc5mk\" (UID: \"3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-pc5mk" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.850682 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6456ccfc99-zhc2h"] Feb 14 11:20:11 crc kubenswrapper[4904]: W0214 11:20:11.877184 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod967e4fe2_b2d0_4526_8d4d_1c0e4c3e6f62.slice/crio-dc63244d8613b57420428a8cd7067373567b9c64a136edb59850e352a4d75880 WatchSource:0}: Error finding container dc63244d8613b57420428a8cd7067373567b9c64a136edb59850e352a4d75880: Status 404 returned error can't find the container with id dc63244d8613b57420428a8cd7067373567b9c64a136edb59850e352a4d75880 Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.922399 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e3e39fe6-7c76-4397-876a-c3a315304a12-console-config\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.922469 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e3e39fe6-7c76-4397-876a-c3a315304a12-console-serving-cert\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.922490 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3e39fe6-7c76-4397-876a-c3a315304a12-trusted-ca-bundle\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.922531 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e3e39fe6-7c76-4397-876a-c3a315304a12-oauth-serving-cert\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.922545 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e3e39fe6-7c76-4397-876a-c3a315304a12-console-oauth-config\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.922566 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bt89\" (UniqueName: \"kubernetes.io/projected/e3e39fe6-7c76-4397-876a-c3a315304a12-kube-api-access-7bt89\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:11 crc kubenswrapper[4904]: I0214 11:20:11.922583 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e3e39fe6-7c76-4397-876a-c3a315304a12-service-ca\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.023289 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e3e39fe6-7c76-4397-876a-c3a315304a12-console-serving-cert\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.023336 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3e39fe6-7c76-4397-876a-c3a315304a12-trusted-ca-bundle\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.023375 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e3e39fe6-7c76-4397-876a-c3a315304a12-oauth-serving-cert\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.023391 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e3e39fe6-7c76-4397-876a-c3a315304a12-console-oauth-config\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.023414 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bt89\" (UniqueName: \"kubernetes.io/projected/e3e39fe6-7c76-4397-876a-c3a315304a12-kube-api-access-7bt89\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.023439 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e3e39fe6-7c76-4397-876a-c3a315304a12-service-ca\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.023484 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e3e39fe6-7c76-4397-876a-c3a315304a12-console-config\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.024337 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e3e39fe6-7c76-4397-876a-c3a315304a12-console-config\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.026394 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e3e39fe6-7c76-4397-876a-c3a315304a12-service-ca\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.027105 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3e39fe6-7c76-4397-876a-c3a315304a12-trusted-ca-bundle\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.027280 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e3e39fe6-7c76-4397-876a-c3a315304a12-console-serving-cert\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.027731 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e3e39fe6-7c76-4397-876a-c3a315304a12-oauth-serving-cert\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.028910 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e3e39fe6-7c76-4397-876a-c3a315304a12-console-oauth-config\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.044542 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bt89\" (UniqueName: \"kubernetes.io/projected/e3e39fe6-7c76-4397-876a-c3a315304a12-kube-api-access-7bt89\") pod \"console-6456ccfc99-zhc2h\" (UID: \"e3e39fe6-7c76-4397-876a-c3a315304a12\") " pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.124454 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/62ca493b-5a80-450e-9e45-8f916f2c05e8-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-tj7n4\" (UID: \"62ca493b-5a80-450e-9e45-8f916f2c05e8\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tj7n4" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.127156 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/62ca493b-5a80-450e-9e45-8f916f2c05e8-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-tj7n4\" (UID: \"62ca493b-5a80-450e-9e45-8f916f2c05e8\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tj7n4" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.163150 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.234157 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-js9s2"] Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.339461 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6456ccfc99-zhc2h"] Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.343544 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-pc5mk\" (UID: \"3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-pc5mk" Feb 14 11:20:12 crc kubenswrapper[4904]: W0214 11:20:12.346098 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3e39fe6_7c76_4397_876a_c3a315304a12.slice/crio-6598be65d78140f07087deab07cd15819840fa0ce21d6e36b653d0bb7d244367 WatchSource:0}: Error finding container 6598be65d78140f07087deab07cd15819840fa0ce21d6e36b653d0bb7d244367: Status 404 returned error can't find the container with id 6598be65d78140f07087deab07cd15819840fa0ce21d6e36b653d0bb7d244367 Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.346589 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-pc5mk\" (UID: \"3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-pc5mk" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.386729 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tj7n4" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.519669 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6456ccfc99-zhc2h" event={"ID":"e3e39fe6-7c76-4397-876a-c3a315304a12","Type":"ContainerStarted","Data":"a594276f6d2d5703abf130bfa51ce129a8bf5011f66cf64cf049536420402c83"} Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.519751 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6456ccfc99-zhc2h" event={"ID":"e3e39fe6-7c76-4397-876a-c3a315304a12","Type":"ContainerStarted","Data":"6598be65d78140f07087deab07cd15819840fa0ce21d6e36b653d0bb7d244367"} Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.521573 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-g5z5r" event={"ID":"967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62","Type":"ContainerStarted","Data":"dc63244d8613b57420428a8cd7067373567b9c64a136edb59850e352a4d75880"} Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.523883 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-pc5mk" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.531646 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-js9s2" event={"ID":"87d08820-f798-415f-a367-e1ec7bfb86d2","Type":"ContainerStarted","Data":"32b92556813c4dc299234cfa48e42a4f0895faa20214d717993f75c70bb3adfd"} Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.539557 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6456ccfc99-zhc2h" podStartSLOduration=1.539535855 podStartE2EDuration="1.539535855s" podCreationTimestamp="2026-02-14 11:20:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:20:12.536819626 +0000 UTC m=+603.349584287" watchObservedRunningTime="2026-02-14 11:20:12.539535855 +0000 UTC m=+603.352300536" Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.582676 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-tj7n4"] Feb 14 11:20:12 crc kubenswrapper[4904]: I0214 11:20:12.781706 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-pc5mk"] Feb 14 11:20:13 crc kubenswrapper[4904]: I0214 11:20:13.537370 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tj7n4" event={"ID":"62ca493b-5a80-450e-9e45-8f916f2c05e8","Type":"ContainerStarted","Data":"f2d70290d93302dea0ecde59a994f0372bd0b401f1ef7d27e0dd8d06b9d6aeec"} Feb 14 11:20:13 crc kubenswrapper[4904]: I0214 11:20:13.538494 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-pc5mk" event={"ID":"3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a","Type":"ContainerStarted","Data":"6f155abb76fc23858b9f5d679d4851409f09b797522b5b668735c88d2458f71b"} Feb 14 11:20:15 crc kubenswrapper[4904]: I0214 11:20:15.566428 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tj7n4" event={"ID":"62ca493b-5a80-450e-9e45-8f916f2c05e8","Type":"ContainerStarted","Data":"57a256783a717834191ed5d0fb008c8db6175030e63dd3ec6c3f3f0337e1cae0"} Feb 14 11:20:15 crc kubenswrapper[4904]: I0214 11:20:15.566886 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tj7n4" Feb 14 11:20:15 crc kubenswrapper[4904]: I0214 11:20:15.574581 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-g5z5r" event={"ID":"967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62","Type":"ContainerStarted","Data":"97918d90e1c4b21e06b62168fb053de0ccf03b5459a74b05a9d24439e3a32dc9"} Feb 14 11:20:15 crc kubenswrapper[4904]: I0214 11:20:15.575482 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-g5z5r" Feb 14 11:20:15 crc kubenswrapper[4904]: I0214 11:20:15.577791 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-js9s2" event={"ID":"87d08820-f798-415f-a367-e1ec7bfb86d2","Type":"ContainerStarted","Data":"fb8feb441155cf8b7ad1af510ef43e951575be7cb38d857fe0bd897881b44e27"} Feb 14 11:20:15 crc kubenswrapper[4904]: I0214 11:20:15.609291 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tj7n4" podStartSLOduration=2.672433319 podStartE2EDuration="4.609271418s" podCreationTimestamp="2026-02-14 11:20:11 +0000 UTC" firstStartedPulling="2026-02-14 11:20:12.598896119 +0000 UTC m=+603.411660780" lastFinishedPulling="2026-02-14 11:20:14.535734218 +0000 UTC m=+605.348498879" observedRunningTime="2026-02-14 11:20:15.591002407 +0000 UTC m=+606.403767068" watchObservedRunningTime="2026-02-14 11:20:15.609271418 +0000 UTC m=+606.422036079" Feb 14 11:20:16 crc kubenswrapper[4904]: I0214 11:20:16.383361 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:20:16 crc kubenswrapper[4904]: I0214 11:20:16.383659 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:20:16 crc kubenswrapper[4904]: I0214 11:20:16.383698 4904 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:20:16 crc kubenswrapper[4904]: I0214 11:20:16.384235 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7bee02c84077b982480197f3573826a9bf5c596ca426336719e5c06a6c3588f2"} pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 14 11:20:16 crc kubenswrapper[4904]: I0214 11:20:16.384299 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" containerID="cri-o://7bee02c84077b982480197f3573826a9bf5c596ca426336719e5c06a6c3588f2" gracePeriod=600 Feb 14 11:20:16 crc kubenswrapper[4904]: I0214 11:20:16.585303 4904 generic.go:334] "Generic (PLEG): container finished" podID="7790d276-2506-4641-8c22-67e6441417f0" containerID="7bee02c84077b982480197f3573826a9bf5c596ca426336719e5c06a6c3588f2" exitCode=0 Feb 14 11:20:16 crc kubenswrapper[4904]: I0214 11:20:16.585357 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerDied","Data":"7bee02c84077b982480197f3573826a9bf5c596ca426336719e5c06a6c3588f2"} Feb 14 11:20:16 crc kubenswrapper[4904]: I0214 11:20:16.585387 4904 scope.go:117] "RemoveContainer" containerID="903ea27dacf6775c4c9f46b3a1c3654b638e3bdce73cd947b4f81a7bea2eca49" Feb 14 11:20:16 crc kubenswrapper[4904]: I0214 11:20:16.586873 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-pc5mk" event={"ID":"3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a","Type":"ContainerStarted","Data":"0ee67a1675d8f94349d464d2137f9f9473c5b22dae5b47ae80cc08caebaa6c30"} Feb 14 11:20:16 crc kubenswrapper[4904]: I0214 11:20:16.604503 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-g5z5r" podStartSLOduration=2.940576248 podStartE2EDuration="5.604485531s" podCreationTimestamp="2026-02-14 11:20:11 +0000 UTC" firstStartedPulling="2026-02-14 11:20:11.88024674 +0000 UTC m=+602.693011401" lastFinishedPulling="2026-02-14 11:20:14.544156023 +0000 UTC m=+605.356920684" observedRunningTime="2026-02-14 11:20:15.650071353 +0000 UTC m=+606.462836014" watchObservedRunningTime="2026-02-14 11:20:16.604485531 +0000 UTC m=+607.417250192" Feb 14 11:20:17 crc kubenswrapper[4904]: I0214 11:20:17.594178 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-js9s2" event={"ID":"87d08820-f798-415f-a367-e1ec7bfb86d2","Type":"ContainerStarted","Data":"66e189aebe7957ee031400ef46996da66a2b9a691c2955795837067f8f62303e"} Feb 14 11:20:17 crc kubenswrapper[4904]: I0214 11:20:17.601396 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"ed0b33120210f8d03c6d331d3186cbee96523120ba158013d88a03648716b22f"} Feb 14 11:20:17 crc kubenswrapper[4904]: I0214 11:20:17.615948 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-pc5mk" podStartSLOduration=3.85315204 podStartE2EDuration="6.615929825s" podCreationTimestamp="2026-02-14 11:20:11 +0000 UTC" firstStartedPulling="2026-02-14 11:20:12.787509919 +0000 UTC m=+603.600274580" lastFinishedPulling="2026-02-14 11:20:15.550287714 +0000 UTC m=+606.363052365" observedRunningTime="2026-02-14 11:20:16.600868956 +0000 UTC m=+607.413633617" watchObservedRunningTime="2026-02-14 11:20:17.615929825 +0000 UTC m=+608.428694496" Feb 14 11:20:17 crc kubenswrapper[4904]: I0214 11:20:17.665919 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58c85c668d-js9s2" podStartSLOduration=2.093092226 podStartE2EDuration="6.665898187s" podCreationTimestamp="2026-02-14 11:20:11 +0000 UTC" firstStartedPulling="2026-02-14 11:20:12.24509764 +0000 UTC m=+603.057862301" lastFinishedPulling="2026-02-14 11:20:16.817903601 +0000 UTC m=+607.630668262" observedRunningTime="2026-02-14 11:20:17.613129964 +0000 UTC m=+608.425894625" watchObservedRunningTime="2026-02-14 11:20:17.665898187 +0000 UTC m=+608.478662868" Feb 14 11:20:21 crc kubenswrapper[4904]: I0214 11:20:21.819308 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-g5z5r" Feb 14 11:20:22 crc kubenswrapper[4904]: I0214 11:20:22.163330 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:22 crc kubenswrapper[4904]: I0214 11:20:22.163409 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:22 crc kubenswrapper[4904]: I0214 11:20:22.169734 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:22 crc kubenswrapper[4904]: I0214 11:20:22.636820 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6456ccfc99-zhc2h" Feb 14 11:20:22 crc kubenswrapper[4904]: I0214 11:20:22.710714 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-7c6cp"] Feb 14 11:20:32 crc kubenswrapper[4904]: I0214 11:20:32.393016 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tj7n4" Feb 14 11:20:44 crc kubenswrapper[4904]: I0214 11:20:44.183785 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t"] Feb 14 11:20:44 crc kubenswrapper[4904]: I0214 11:20:44.185435 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t" Feb 14 11:20:44 crc kubenswrapper[4904]: I0214 11:20:44.187213 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 14 11:20:44 crc kubenswrapper[4904]: I0214 11:20:44.198137 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t"] Feb 14 11:20:44 crc kubenswrapper[4904]: I0214 11:20:44.287447 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4eac63f-c406-48a9-ac8b-6150cece2deb-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t\" (UID: \"b4eac63f-c406-48a9-ac8b-6150cece2deb\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t" Feb 14 11:20:44 crc kubenswrapper[4904]: I0214 11:20:44.287526 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4eac63f-c406-48a9-ac8b-6150cece2deb-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t\" (UID: \"b4eac63f-c406-48a9-ac8b-6150cece2deb\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t" Feb 14 11:20:44 crc kubenswrapper[4904]: I0214 11:20:44.287553 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9l4g\" (UniqueName: \"kubernetes.io/projected/b4eac63f-c406-48a9-ac8b-6150cece2deb-kube-api-access-x9l4g\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t\" (UID: \"b4eac63f-c406-48a9-ac8b-6150cece2deb\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t" Feb 14 11:20:44 crc kubenswrapper[4904]: I0214 11:20:44.388957 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4eac63f-c406-48a9-ac8b-6150cece2deb-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t\" (UID: \"b4eac63f-c406-48a9-ac8b-6150cece2deb\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t" Feb 14 11:20:44 crc kubenswrapper[4904]: I0214 11:20:44.389003 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9l4g\" (UniqueName: \"kubernetes.io/projected/b4eac63f-c406-48a9-ac8b-6150cece2deb-kube-api-access-x9l4g\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t\" (UID: \"b4eac63f-c406-48a9-ac8b-6150cece2deb\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t" Feb 14 11:20:44 crc kubenswrapper[4904]: I0214 11:20:44.389092 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4eac63f-c406-48a9-ac8b-6150cece2deb-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t\" (UID: \"b4eac63f-c406-48a9-ac8b-6150cece2deb\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t" Feb 14 11:20:44 crc kubenswrapper[4904]: I0214 11:20:44.389511 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4eac63f-c406-48a9-ac8b-6150cece2deb-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t\" (UID: \"b4eac63f-c406-48a9-ac8b-6150cece2deb\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t" Feb 14 11:20:44 crc kubenswrapper[4904]: I0214 11:20:44.389743 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4eac63f-c406-48a9-ac8b-6150cece2deb-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t\" (UID: \"b4eac63f-c406-48a9-ac8b-6150cece2deb\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t" Feb 14 11:20:44 crc kubenswrapper[4904]: I0214 11:20:44.406752 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9l4g\" (UniqueName: \"kubernetes.io/projected/b4eac63f-c406-48a9-ac8b-6150cece2deb-kube-api-access-x9l4g\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t\" (UID: \"b4eac63f-c406-48a9-ac8b-6150cece2deb\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t" Feb 14 11:20:44 crc kubenswrapper[4904]: I0214 11:20:44.502738 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t" Feb 14 11:20:44 crc kubenswrapper[4904]: I0214 11:20:44.919760 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t"] Feb 14 11:20:45 crc kubenswrapper[4904]: I0214 11:20:45.754366 4904 generic.go:334] "Generic (PLEG): container finished" podID="b4eac63f-c406-48a9-ac8b-6150cece2deb" containerID="d66c9005b75599233b7bb2dcad4c75b4a26a9b071b35133b8b6babc8477d655d" exitCode=0 Feb 14 11:20:45 crc kubenswrapper[4904]: I0214 11:20:45.754420 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t" event={"ID":"b4eac63f-c406-48a9-ac8b-6150cece2deb","Type":"ContainerDied","Data":"d66c9005b75599233b7bb2dcad4c75b4a26a9b071b35133b8b6babc8477d655d"} Feb 14 11:20:45 crc kubenswrapper[4904]: I0214 11:20:45.754626 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t" event={"ID":"b4eac63f-c406-48a9-ac8b-6150cece2deb","Type":"ContainerStarted","Data":"8fd381a8239cc8db7ab50ebd2a55c72e8d0fd1da722e5ba23ff534953cb07d88"} Feb 14 11:20:46 crc kubenswrapper[4904]: I0214 11:20:46.768485 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t" event={"ID":"b4eac63f-c406-48a9-ac8b-6150cece2deb","Type":"ContainerStarted","Data":"4214ae88eff34b537a3932e1a4fd62b76f4656d2ecbf251e87ecee17cbfeb17d"} Feb 14 11:20:47 crc kubenswrapper[4904]: I0214 11:20:47.752260 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-7c6cp" podUID="261ad318-57b3-48ca-8dac-e8bd53528694" containerName="console" containerID="cri-o://0ceb5bec8f16f4e10c8554e8100955c5da044ec2434441963b8f51cf7c0d110a" gracePeriod=15 Feb 14 11:20:47 crc kubenswrapper[4904]: I0214 11:20:47.784620 4904 generic.go:334] "Generic (PLEG): container finished" podID="b4eac63f-c406-48a9-ac8b-6150cece2deb" containerID="4214ae88eff34b537a3932e1a4fd62b76f4656d2ecbf251e87ecee17cbfeb17d" exitCode=0 Feb 14 11:20:47 crc kubenswrapper[4904]: I0214 11:20:47.784672 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t" event={"ID":"b4eac63f-c406-48a9-ac8b-6150cece2deb","Type":"ContainerDied","Data":"4214ae88eff34b537a3932e1a4fd62b76f4656d2ecbf251e87ecee17cbfeb17d"} Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.142252 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-7c6cp_261ad318-57b3-48ca-8dac-e8bd53528694/console/0.log" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.142550 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.238145 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/261ad318-57b3-48ca-8dac-e8bd53528694-console-serving-cert\") pod \"261ad318-57b3-48ca-8dac-e8bd53528694\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.238192 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-oauth-serving-cert\") pod \"261ad318-57b3-48ca-8dac-e8bd53528694\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.238287 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-console-config\") pod \"261ad318-57b3-48ca-8dac-e8bd53528694\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.238317 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-trusted-ca-bundle\") pod \"261ad318-57b3-48ca-8dac-e8bd53528694\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.238349 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/261ad318-57b3-48ca-8dac-e8bd53528694-console-oauth-config\") pod \"261ad318-57b3-48ca-8dac-e8bd53528694\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.238377 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-service-ca\") pod \"261ad318-57b3-48ca-8dac-e8bd53528694\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.238397 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pb5wr\" (UniqueName: \"kubernetes.io/projected/261ad318-57b3-48ca-8dac-e8bd53528694-kube-api-access-pb5wr\") pod \"261ad318-57b3-48ca-8dac-e8bd53528694\" (UID: \"261ad318-57b3-48ca-8dac-e8bd53528694\") " Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.239758 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-console-config" (OuterVolumeSpecName: "console-config") pod "261ad318-57b3-48ca-8dac-e8bd53528694" (UID: "261ad318-57b3-48ca-8dac-e8bd53528694"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.240583 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-service-ca" (OuterVolumeSpecName: "service-ca") pod "261ad318-57b3-48ca-8dac-e8bd53528694" (UID: "261ad318-57b3-48ca-8dac-e8bd53528694"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.240632 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "261ad318-57b3-48ca-8dac-e8bd53528694" (UID: "261ad318-57b3-48ca-8dac-e8bd53528694"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.241699 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "261ad318-57b3-48ca-8dac-e8bd53528694" (UID: "261ad318-57b3-48ca-8dac-e8bd53528694"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.244324 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/261ad318-57b3-48ca-8dac-e8bd53528694-kube-api-access-pb5wr" (OuterVolumeSpecName: "kube-api-access-pb5wr") pod "261ad318-57b3-48ca-8dac-e8bd53528694" (UID: "261ad318-57b3-48ca-8dac-e8bd53528694"). InnerVolumeSpecName "kube-api-access-pb5wr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.246476 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/261ad318-57b3-48ca-8dac-e8bd53528694-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "261ad318-57b3-48ca-8dac-e8bd53528694" (UID: "261ad318-57b3-48ca-8dac-e8bd53528694"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.247107 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/261ad318-57b3-48ca-8dac-e8bd53528694-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "261ad318-57b3-48ca-8dac-e8bd53528694" (UID: "261ad318-57b3-48ca-8dac-e8bd53528694"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.339723 4904 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-service-ca\") on node \"crc\" DevicePath \"\"" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.339994 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pb5wr\" (UniqueName: \"kubernetes.io/projected/261ad318-57b3-48ca-8dac-e8bd53528694-kube-api-access-pb5wr\") on node \"crc\" DevicePath \"\"" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.340061 4904 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/261ad318-57b3-48ca-8dac-e8bd53528694-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.340125 4904 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.340183 4904 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-console-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.340237 4904 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/261ad318-57b3-48ca-8dac-e8bd53528694-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.340294 4904 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/261ad318-57b3-48ca-8dac-e8bd53528694-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.793335 4904 generic.go:334] "Generic (PLEG): container finished" podID="b4eac63f-c406-48a9-ac8b-6150cece2deb" containerID="d5022d292af241e14a842b607203dfb7fa6b466068e0f6893c78ee51c99785e9" exitCode=0 Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.793373 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t" event={"ID":"b4eac63f-c406-48a9-ac8b-6150cece2deb","Type":"ContainerDied","Data":"d5022d292af241e14a842b607203dfb7fa6b466068e0f6893c78ee51c99785e9"} Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.795502 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-7c6cp_261ad318-57b3-48ca-8dac-e8bd53528694/console/0.log" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.795538 4904 generic.go:334] "Generic (PLEG): container finished" podID="261ad318-57b3-48ca-8dac-e8bd53528694" containerID="0ceb5bec8f16f4e10c8554e8100955c5da044ec2434441963b8f51cf7c0d110a" exitCode=2 Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.795562 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7c6cp" event={"ID":"261ad318-57b3-48ca-8dac-e8bd53528694","Type":"ContainerDied","Data":"0ceb5bec8f16f4e10c8554e8100955c5da044ec2434441963b8f51cf7c0d110a"} Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.795583 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7c6cp" event={"ID":"261ad318-57b3-48ca-8dac-e8bd53528694","Type":"ContainerDied","Data":"7fccd0ed3d7eaaf404719a88d37ef0cf98a037d712dbc05f1db3e966cd5bcbfd"} Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.795602 4904 scope.go:117] "RemoveContainer" containerID="0ceb5bec8f16f4e10c8554e8100955c5da044ec2434441963b8f51cf7c0d110a" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.795738 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7c6cp" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.811593 4904 scope.go:117] "RemoveContainer" containerID="0ceb5bec8f16f4e10c8554e8100955c5da044ec2434441963b8f51cf7c0d110a" Feb 14 11:20:48 crc kubenswrapper[4904]: E0214 11:20:48.811958 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ceb5bec8f16f4e10c8554e8100955c5da044ec2434441963b8f51cf7c0d110a\": container with ID starting with 0ceb5bec8f16f4e10c8554e8100955c5da044ec2434441963b8f51cf7c0d110a not found: ID does not exist" containerID="0ceb5bec8f16f4e10c8554e8100955c5da044ec2434441963b8f51cf7c0d110a" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.811987 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ceb5bec8f16f4e10c8554e8100955c5da044ec2434441963b8f51cf7c0d110a"} err="failed to get container status \"0ceb5bec8f16f4e10c8554e8100955c5da044ec2434441963b8f51cf7c0d110a\": rpc error: code = NotFound desc = could not find container \"0ceb5bec8f16f4e10c8554e8100955c5da044ec2434441963b8f51cf7c0d110a\": container with ID starting with 0ceb5bec8f16f4e10c8554e8100955c5da044ec2434441963b8f51cf7c0d110a not found: ID does not exist" Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.827024 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-7c6cp"] Feb 14 11:20:48 crc kubenswrapper[4904]: I0214 11:20:48.831007 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-7c6cp"] Feb 14 11:20:49 crc kubenswrapper[4904]: I0214 11:20:49.850014 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="261ad318-57b3-48ca-8dac-e8bd53528694" path="/var/lib/kubelet/pods/261ad318-57b3-48ca-8dac-e8bd53528694/volumes" Feb 14 11:20:50 crc kubenswrapper[4904]: I0214 11:20:50.048101 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t" Feb 14 11:20:50 crc kubenswrapper[4904]: I0214 11:20:50.168053 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9l4g\" (UniqueName: \"kubernetes.io/projected/b4eac63f-c406-48a9-ac8b-6150cece2deb-kube-api-access-x9l4g\") pod \"b4eac63f-c406-48a9-ac8b-6150cece2deb\" (UID: \"b4eac63f-c406-48a9-ac8b-6150cece2deb\") " Feb 14 11:20:50 crc kubenswrapper[4904]: I0214 11:20:50.168121 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4eac63f-c406-48a9-ac8b-6150cece2deb-util\") pod \"b4eac63f-c406-48a9-ac8b-6150cece2deb\" (UID: \"b4eac63f-c406-48a9-ac8b-6150cece2deb\") " Feb 14 11:20:50 crc kubenswrapper[4904]: I0214 11:20:50.168170 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4eac63f-c406-48a9-ac8b-6150cece2deb-bundle\") pod \"b4eac63f-c406-48a9-ac8b-6150cece2deb\" (UID: \"b4eac63f-c406-48a9-ac8b-6150cece2deb\") " Feb 14 11:20:50 crc kubenswrapper[4904]: I0214 11:20:50.169172 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4eac63f-c406-48a9-ac8b-6150cece2deb-bundle" (OuterVolumeSpecName: "bundle") pod "b4eac63f-c406-48a9-ac8b-6150cece2deb" (UID: "b4eac63f-c406-48a9-ac8b-6150cece2deb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:20:50 crc kubenswrapper[4904]: I0214 11:20:50.176988 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4eac63f-c406-48a9-ac8b-6150cece2deb-kube-api-access-x9l4g" (OuterVolumeSpecName: "kube-api-access-x9l4g") pod "b4eac63f-c406-48a9-ac8b-6150cece2deb" (UID: "b4eac63f-c406-48a9-ac8b-6150cece2deb"). InnerVolumeSpecName "kube-api-access-x9l4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:20:50 crc kubenswrapper[4904]: I0214 11:20:50.183158 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4eac63f-c406-48a9-ac8b-6150cece2deb-util" (OuterVolumeSpecName: "util") pod "b4eac63f-c406-48a9-ac8b-6150cece2deb" (UID: "b4eac63f-c406-48a9-ac8b-6150cece2deb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:20:50 crc kubenswrapper[4904]: I0214 11:20:50.269323 4904 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4eac63f-c406-48a9-ac8b-6150cece2deb-util\") on node \"crc\" DevicePath \"\"" Feb 14 11:20:50 crc kubenswrapper[4904]: I0214 11:20:50.269369 4904 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4eac63f-c406-48a9-ac8b-6150cece2deb-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:20:50 crc kubenswrapper[4904]: I0214 11:20:50.269384 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9l4g\" (UniqueName: \"kubernetes.io/projected/b4eac63f-c406-48a9-ac8b-6150cece2deb-kube-api-access-x9l4g\") on node \"crc\" DevicePath \"\"" Feb 14 11:20:50 crc kubenswrapper[4904]: I0214 11:20:50.809337 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t" event={"ID":"b4eac63f-c406-48a9-ac8b-6150cece2deb","Type":"ContainerDied","Data":"8fd381a8239cc8db7ab50ebd2a55c72e8d0fd1da722e5ba23ff534953cb07d88"} Feb 14 11:20:50 crc kubenswrapper[4904]: I0214 11:20:50.809378 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fd381a8239cc8db7ab50ebd2a55c72e8d0fd1da722e5ba23ff534953cb07d88" Feb 14 11:20:50 crc kubenswrapper[4904]: I0214 11:20:50.809387 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t" Feb 14 11:20:59 crc kubenswrapper[4904]: I0214 11:20:59.984439 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7b865f4cbf-8thhc"] Feb 14 11:20:59 crc kubenswrapper[4904]: E0214 11:20:59.985129 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4eac63f-c406-48a9-ac8b-6150cece2deb" containerName="extract" Feb 14 11:20:59 crc kubenswrapper[4904]: I0214 11:20:59.985140 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4eac63f-c406-48a9-ac8b-6150cece2deb" containerName="extract" Feb 14 11:20:59 crc kubenswrapper[4904]: E0214 11:20:59.985151 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4eac63f-c406-48a9-ac8b-6150cece2deb" containerName="util" Feb 14 11:20:59 crc kubenswrapper[4904]: I0214 11:20:59.985156 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4eac63f-c406-48a9-ac8b-6150cece2deb" containerName="util" Feb 14 11:20:59 crc kubenswrapper[4904]: E0214 11:20:59.985174 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4eac63f-c406-48a9-ac8b-6150cece2deb" containerName="pull" Feb 14 11:20:59 crc kubenswrapper[4904]: I0214 11:20:59.985180 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4eac63f-c406-48a9-ac8b-6150cece2deb" containerName="pull" Feb 14 11:20:59 crc kubenswrapper[4904]: E0214 11:20:59.985190 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="261ad318-57b3-48ca-8dac-e8bd53528694" containerName="console" Feb 14 11:20:59 crc kubenswrapper[4904]: I0214 11:20:59.985195 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="261ad318-57b3-48ca-8dac-e8bd53528694" containerName="console" Feb 14 11:20:59 crc kubenswrapper[4904]: I0214 11:20:59.985278 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="261ad318-57b3-48ca-8dac-e8bd53528694" containerName="console" Feb 14 11:20:59 crc kubenswrapper[4904]: I0214 11:20:59.985287 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4eac63f-c406-48a9-ac8b-6150cece2deb" containerName="extract" Feb 14 11:20:59 crc kubenswrapper[4904]: I0214 11:20:59.985688 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7b865f4cbf-8thhc" Feb 14 11:20:59 crc kubenswrapper[4904]: I0214 11:20:59.988266 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 14 11:20:59 crc kubenswrapper[4904]: I0214 11:20:59.988458 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 14 11:20:59 crc kubenswrapper[4904]: I0214 11:20:59.988574 4904 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 14 11:20:59 crc kubenswrapper[4904]: I0214 11:20:59.988696 4904 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-gz9fr" Feb 14 11:20:59 crc kubenswrapper[4904]: I0214 11:20:59.995622 4904 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.007706 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7b865f4cbf-8thhc"] Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.088903 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/678c1429-9ad4-44cd-8f86-29bd34395c5b-apiservice-cert\") pod \"metallb-operator-controller-manager-7b865f4cbf-8thhc\" (UID: \"678c1429-9ad4-44cd-8f86-29bd34395c5b\") " pod="metallb-system/metallb-operator-controller-manager-7b865f4cbf-8thhc" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.088963 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4jpn\" (UniqueName: \"kubernetes.io/projected/678c1429-9ad4-44cd-8f86-29bd34395c5b-kube-api-access-l4jpn\") pod \"metallb-operator-controller-manager-7b865f4cbf-8thhc\" (UID: \"678c1429-9ad4-44cd-8f86-29bd34395c5b\") " pod="metallb-system/metallb-operator-controller-manager-7b865f4cbf-8thhc" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.088999 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/678c1429-9ad4-44cd-8f86-29bd34395c5b-webhook-cert\") pod \"metallb-operator-controller-manager-7b865f4cbf-8thhc\" (UID: \"678c1429-9ad4-44cd-8f86-29bd34395c5b\") " pod="metallb-system/metallb-operator-controller-manager-7b865f4cbf-8thhc" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.189779 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/678c1429-9ad4-44cd-8f86-29bd34395c5b-apiservice-cert\") pod \"metallb-operator-controller-manager-7b865f4cbf-8thhc\" (UID: \"678c1429-9ad4-44cd-8f86-29bd34395c5b\") " pod="metallb-system/metallb-operator-controller-manager-7b865f4cbf-8thhc" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.189826 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4jpn\" (UniqueName: \"kubernetes.io/projected/678c1429-9ad4-44cd-8f86-29bd34395c5b-kube-api-access-l4jpn\") pod \"metallb-operator-controller-manager-7b865f4cbf-8thhc\" (UID: \"678c1429-9ad4-44cd-8f86-29bd34395c5b\") " pod="metallb-system/metallb-operator-controller-manager-7b865f4cbf-8thhc" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.189876 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/678c1429-9ad4-44cd-8f86-29bd34395c5b-webhook-cert\") pod \"metallb-operator-controller-manager-7b865f4cbf-8thhc\" (UID: \"678c1429-9ad4-44cd-8f86-29bd34395c5b\") " pod="metallb-system/metallb-operator-controller-manager-7b865f4cbf-8thhc" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.195245 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/678c1429-9ad4-44cd-8f86-29bd34395c5b-webhook-cert\") pod \"metallb-operator-controller-manager-7b865f4cbf-8thhc\" (UID: \"678c1429-9ad4-44cd-8f86-29bd34395c5b\") " pod="metallb-system/metallb-operator-controller-manager-7b865f4cbf-8thhc" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.196063 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/678c1429-9ad4-44cd-8f86-29bd34395c5b-apiservice-cert\") pod \"metallb-operator-controller-manager-7b865f4cbf-8thhc\" (UID: \"678c1429-9ad4-44cd-8f86-29bd34395c5b\") " pod="metallb-system/metallb-operator-controller-manager-7b865f4cbf-8thhc" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.206551 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4jpn\" (UniqueName: \"kubernetes.io/projected/678c1429-9ad4-44cd-8f86-29bd34395c5b-kube-api-access-l4jpn\") pod \"metallb-operator-controller-manager-7b865f4cbf-8thhc\" (UID: \"678c1429-9ad4-44cd-8f86-29bd34395c5b\") " pod="metallb-system/metallb-operator-controller-manager-7b865f4cbf-8thhc" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.303491 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7b865f4cbf-8thhc" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.401128 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-d55fb6f5-cljdr"] Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.401872 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-d55fb6f5-cljdr" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.406296 4904 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.406385 4904 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.406628 4904 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-lsrqz" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.437618 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-d55fb6f5-cljdr"] Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.496771 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svtf9\" (UniqueName: \"kubernetes.io/projected/481cda22-3fb4-4c12-a4ed-29b316a69915-kube-api-access-svtf9\") pod \"metallb-operator-webhook-server-d55fb6f5-cljdr\" (UID: \"481cda22-3fb4-4c12-a4ed-29b316a69915\") " pod="metallb-system/metallb-operator-webhook-server-d55fb6f5-cljdr" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.496826 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/481cda22-3fb4-4c12-a4ed-29b316a69915-webhook-cert\") pod \"metallb-operator-webhook-server-d55fb6f5-cljdr\" (UID: \"481cda22-3fb4-4c12-a4ed-29b316a69915\") " pod="metallb-system/metallb-operator-webhook-server-d55fb6f5-cljdr" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.496909 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/481cda22-3fb4-4c12-a4ed-29b316a69915-apiservice-cert\") pod \"metallb-operator-webhook-server-d55fb6f5-cljdr\" (UID: \"481cda22-3fb4-4c12-a4ed-29b316a69915\") " pod="metallb-system/metallb-operator-webhook-server-d55fb6f5-cljdr" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.597588 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/481cda22-3fb4-4c12-a4ed-29b316a69915-apiservice-cert\") pod \"metallb-operator-webhook-server-d55fb6f5-cljdr\" (UID: \"481cda22-3fb4-4c12-a4ed-29b316a69915\") " pod="metallb-system/metallb-operator-webhook-server-d55fb6f5-cljdr" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.597646 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svtf9\" (UniqueName: \"kubernetes.io/projected/481cda22-3fb4-4c12-a4ed-29b316a69915-kube-api-access-svtf9\") pod \"metallb-operator-webhook-server-d55fb6f5-cljdr\" (UID: \"481cda22-3fb4-4c12-a4ed-29b316a69915\") " pod="metallb-system/metallb-operator-webhook-server-d55fb6f5-cljdr" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.597679 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/481cda22-3fb4-4c12-a4ed-29b316a69915-webhook-cert\") pod \"metallb-operator-webhook-server-d55fb6f5-cljdr\" (UID: \"481cda22-3fb4-4c12-a4ed-29b316a69915\") " pod="metallb-system/metallb-operator-webhook-server-d55fb6f5-cljdr" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.608543 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/481cda22-3fb4-4c12-a4ed-29b316a69915-apiservice-cert\") pod \"metallb-operator-webhook-server-d55fb6f5-cljdr\" (UID: \"481cda22-3fb4-4c12-a4ed-29b316a69915\") " pod="metallb-system/metallb-operator-webhook-server-d55fb6f5-cljdr" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.616401 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/481cda22-3fb4-4c12-a4ed-29b316a69915-webhook-cert\") pod \"metallb-operator-webhook-server-d55fb6f5-cljdr\" (UID: \"481cda22-3fb4-4c12-a4ed-29b316a69915\") " pod="metallb-system/metallb-operator-webhook-server-d55fb6f5-cljdr" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.625580 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svtf9\" (UniqueName: \"kubernetes.io/projected/481cda22-3fb4-4c12-a4ed-29b316a69915-kube-api-access-svtf9\") pod \"metallb-operator-webhook-server-d55fb6f5-cljdr\" (UID: \"481cda22-3fb4-4c12-a4ed-29b316a69915\") " pod="metallb-system/metallb-operator-webhook-server-d55fb6f5-cljdr" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.717945 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7b865f4cbf-8thhc"] Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.738451 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-d55fb6f5-cljdr" Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.879902 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7b865f4cbf-8thhc" event={"ID":"678c1429-9ad4-44cd-8f86-29bd34395c5b","Type":"ContainerStarted","Data":"d6957abf6c9f6b7c2a13572ddcae03756eb06a30a55d69490a08d2cdd55f338e"} Feb 14 11:21:00 crc kubenswrapper[4904]: I0214 11:21:00.986842 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-d55fb6f5-cljdr"] Feb 14 11:21:01 crc kubenswrapper[4904]: I0214 11:21:01.886794 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-d55fb6f5-cljdr" event={"ID":"481cda22-3fb4-4c12-a4ed-29b316a69915","Type":"ContainerStarted","Data":"d9b59ba0ff8a859a7875a8d93fd9ce8b5b0dc0b0d7137ce0971e801a99b5132d"} Feb 14 11:21:04 crc kubenswrapper[4904]: I0214 11:21:04.904608 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7b865f4cbf-8thhc" event={"ID":"678c1429-9ad4-44cd-8f86-29bd34395c5b","Type":"ContainerStarted","Data":"9454945a828ae013fe7f944bd58ecd81663e4188d0883fd1e4faf51978d480e5"} Feb 14 11:21:04 crc kubenswrapper[4904]: I0214 11:21:04.905294 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7b865f4cbf-8thhc" Feb 14 11:21:04 crc kubenswrapper[4904]: I0214 11:21:04.929530 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7b865f4cbf-8thhc" podStartSLOduration=2.7637673510000003 podStartE2EDuration="5.929511862s" podCreationTimestamp="2026-02-14 11:20:59 +0000 UTC" firstStartedPulling="2026-02-14 11:21:00.722141191 +0000 UTC m=+651.534905852" lastFinishedPulling="2026-02-14 11:21:03.887885702 +0000 UTC m=+654.700650363" observedRunningTime="2026-02-14 11:21:04.927034381 +0000 UTC m=+655.739799062" watchObservedRunningTime="2026-02-14 11:21:04.929511862 +0000 UTC m=+655.742276533" Feb 14 11:21:06 crc kubenswrapper[4904]: I0214 11:21:06.917930 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-d55fb6f5-cljdr" event={"ID":"481cda22-3fb4-4c12-a4ed-29b316a69915","Type":"ContainerStarted","Data":"a9758d6fd8e6e312c9c289b7c46a42aab306a08773b610fd9b490b004396dc23"} Feb 14 11:21:06 crc kubenswrapper[4904]: I0214 11:21:06.918456 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-d55fb6f5-cljdr" Feb 14 11:21:20 crc kubenswrapper[4904]: I0214 11:21:20.772717 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-d55fb6f5-cljdr" Feb 14 11:21:20 crc kubenswrapper[4904]: I0214 11:21:20.796161 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-d55fb6f5-cljdr" podStartSLOduration=15.302330511 podStartE2EDuration="20.796140692s" podCreationTimestamp="2026-02-14 11:21:00 +0000 UTC" firstStartedPulling="2026-02-14 11:21:00.994520529 +0000 UTC m=+651.807285180" lastFinishedPulling="2026-02-14 11:21:06.48833069 +0000 UTC m=+657.301095361" observedRunningTime="2026-02-14 11:21:06.942375628 +0000 UTC m=+657.755140289" watchObservedRunningTime="2026-02-14 11:21:20.796140692 +0000 UTC m=+671.608905353" Feb 14 11:21:40 crc kubenswrapper[4904]: I0214 11:21:40.306940 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7b865f4cbf-8thhc" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.029606 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-jzq7k"] Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.030346 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jzq7k" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.033558 4904 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-j8vl5" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.033689 4904 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.036329 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-x6sv9"] Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.038392 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.046682 4904 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.046974 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.112539 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-jzq7k"] Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.142901 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-zmg22"] Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.143700 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-zmg22" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.145913 4904 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.145947 4904 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.146899 4904 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-2dgbv" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.147093 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.182765 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-69bbfbf88f-bf7zh"] Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.183601 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-bf7zh" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.186804 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/bdf867d3-84cd-4b7c-8e49-97ba408a1689-frr-conf\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.186860 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7j2ml\" (UniqueName: \"kubernetes.io/projected/b15d7113-f0f8-4033-a293-fc94960799c4-kube-api-access-7j2ml\") pod \"frr-k8s-webhook-server-78b44bf5bb-jzq7k\" (UID: \"b15d7113-f0f8-4033-a293-fc94960799c4\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jzq7k" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.187096 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/bdf867d3-84cd-4b7c-8e49-97ba408a1689-frr-startup\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.187126 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/bdf867d3-84cd-4b7c-8e49-97ba408a1689-metrics\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.187153 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/bdf867d3-84cd-4b7c-8e49-97ba408a1689-reloader\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.187175 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nllsw\" (UniqueName: \"kubernetes.io/projected/bdf867d3-84cd-4b7c-8e49-97ba408a1689-kube-api-access-nllsw\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.187233 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/bdf867d3-84cd-4b7c-8e49-97ba408a1689-frr-sockets\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.187252 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b15d7113-f0f8-4033-a293-fc94960799c4-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-jzq7k\" (UID: \"b15d7113-f0f8-4033-a293-fc94960799c4\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jzq7k" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.187273 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bdf867d3-84cd-4b7c-8e49-97ba408a1689-metrics-certs\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.190221 4904 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.209430 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-bf7zh"] Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.288249 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5k76\" (UniqueName: \"kubernetes.io/projected/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-kube-api-access-r5k76\") pod \"speaker-zmg22\" (UID: \"90c810e2-d81e-44b0-bf0c-ec3ba2776c55\") " pod="metallb-system/speaker-zmg22" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.288312 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7j2ml\" (UniqueName: \"kubernetes.io/projected/b15d7113-f0f8-4033-a293-fc94960799c4-kube-api-access-7j2ml\") pod \"frr-k8s-webhook-server-78b44bf5bb-jzq7k\" (UID: \"b15d7113-f0f8-4033-a293-fc94960799c4\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jzq7k" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.288339 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-memberlist\") pod \"speaker-zmg22\" (UID: \"90c810e2-d81e-44b0-bf0c-ec3ba2776c55\") " pod="metallb-system/speaker-zmg22" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.288379 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/bdf867d3-84cd-4b7c-8e49-97ba408a1689-frr-startup\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.288401 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/bdf867d3-84cd-4b7c-8e49-97ba408a1689-metrics\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.288429 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/bdf867d3-84cd-4b7c-8e49-97ba408a1689-reloader\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.288460 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nllsw\" (UniqueName: \"kubernetes.io/projected/bdf867d3-84cd-4b7c-8e49-97ba408a1689-kube-api-access-nllsw\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.288488 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/bdf867d3-84cd-4b7c-8e49-97ba408a1689-frr-sockets\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.288506 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b15d7113-f0f8-4033-a293-fc94960799c4-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-jzq7k\" (UID: \"b15d7113-f0f8-4033-a293-fc94960799c4\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jzq7k" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.288538 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bdf867d3-84cd-4b7c-8e49-97ba408a1689-metrics-certs\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.288580 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls4gd\" (UniqueName: \"kubernetes.io/projected/ed02a813-b989-405f-87d8-6b4ed3539671-kube-api-access-ls4gd\") pod \"controller-69bbfbf88f-bf7zh\" (UID: \"ed02a813-b989-405f-87d8-6b4ed3539671\") " pod="metallb-system/controller-69bbfbf88f-bf7zh" Feb 14 11:21:41 crc kubenswrapper[4904]: E0214 11:21:41.288614 4904 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.288631 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ed02a813-b989-405f-87d8-6b4ed3539671-cert\") pod \"controller-69bbfbf88f-bf7zh\" (UID: \"ed02a813-b989-405f-87d8-6b4ed3539671\") " pod="metallb-system/controller-69bbfbf88f-bf7zh" Feb 14 11:21:41 crc kubenswrapper[4904]: E0214 11:21:41.288668 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b15d7113-f0f8-4033-a293-fc94960799c4-cert podName:b15d7113-f0f8-4033-a293-fc94960799c4 nodeName:}" failed. No retries permitted until 2026-02-14 11:21:41.788653414 +0000 UTC m=+692.601418075 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b15d7113-f0f8-4033-a293-fc94960799c4-cert") pod "frr-k8s-webhook-server-78b44bf5bb-jzq7k" (UID: "b15d7113-f0f8-4033-a293-fc94960799c4") : secret "frr-k8s-webhook-server-cert" not found Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.288692 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-metallb-excludel2\") pod \"speaker-zmg22\" (UID: \"90c810e2-d81e-44b0-bf0c-ec3ba2776c55\") " pod="metallb-system/speaker-zmg22" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.288721 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ed02a813-b989-405f-87d8-6b4ed3539671-metrics-certs\") pod \"controller-69bbfbf88f-bf7zh\" (UID: \"ed02a813-b989-405f-87d8-6b4ed3539671\") " pod="metallb-system/controller-69bbfbf88f-bf7zh" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.288754 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-metrics-certs\") pod \"speaker-zmg22\" (UID: \"90c810e2-d81e-44b0-bf0c-ec3ba2776c55\") " pod="metallb-system/speaker-zmg22" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.288794 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/bdf867d3-84cd-4b7c-8e49-97ba408a1689-frr-conf\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.289185 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/bdf867d3-84cd-4b7c-8e49-97ba408a1689-metrics\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.289237 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/bdf867d3-84cd-4b7c-8e49-97ba408a1689-frr-sockets\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.289308 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/bdf867d3-84cd-4b7c-8e49-97ba408a1689-reloader\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.289401 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/bdf867d3-84cd-4b7c-8e49-97ba408a1689-frr-conf\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.289942 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/bdf867d3-84cd-4b7c-8e49-97ba408a1689-frr-startup\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.295403 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bdf867d3-84cd-4b7c-8e49-97ba408a1689-metrics-certs\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.310113 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nllsw\" (UniqueName: \"kubernetes.io/projected/bdf867d3-84cd-4b7c-8e49-97ba408a1689-kube-api-access-nllsw\") pod \"frr-k8s-x6sv9\" (UID: \"bdf867d3-84cd-4b7c-8e49-97ba408a1689\") " pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.315965 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7j2ml\" (UniqueName: \"kubernetes.io/projected/b15d7113-f0f8-4033-a293-fc94960799c4-kube-api-access-7j2ml\") pod \"frr-k8s-webhook-server-78b44bf5bb-jzq7k\" (UID: \"b15d7113-f0f8-4033-a293-fc94960799c4\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jzq7k" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.372110 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.392676 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5k76\" (UniqueName: \"kubernetes.io/projected/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-kube-api-access-r5k76\") pod \"speaker-zmg22\" (UID: \"90c810e2-d81e-44b0-bf0c-ec3ba2776c55\") " pod="metallb-system/speaker-zmg22" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.392760 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-memberlist\") pod \"speaker-zmg22\" (UID: \"90c810e2-d81e-44b0-bf0c-ec3ba2776c55\") " pod="metallb-system/speaker-zmg22" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.392877 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls4gd\" (UniqueName: \"kubernetes.io/projected/ed02a813-b989-405f-87d8-6b4ed3539671-kube-api-access-ls4gd\") pod \"controller-69bbfbf88f-bf7zh\" (UID: \"ed02a813-b989-405f-87d8-6b4ed3539671\") " pod="metallb-system/controller-69bbfbf88f-bf7zh" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.392912 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ed02a813-b989-405f-87d8-6b4ed3539671-cert\") pod \"controller-69bbfbf88f-bf7zh\" (UID: \"ed02a813-b989-405f-87d8-6b4ed3539671\") " pod="metallb-system/controller-69bbfbf88f-bf7zh" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.392942 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-metallb-excludel2\") pod \"speaker-zmg22\" (UID: \"90c810e2-d81e-44b0-bf0c-ec3ba2776c55\") " pod="metallb-system/speaker-zmg22" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.392965 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ed02a813-b989-405f-87d8-6b4ed3539671-metrics-certs\") pod \"controller-69bbfbf88f-bf7zh\" (UID: \"ed02a813-b989-405f-87d8-6b4ed3539671\") " pod="metallb-system/controller-69bbfbf88f-bf7zh" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.392990 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-metrics-certs\") pod \"speaker-zmg22\" (UID: \"90c810e2-d81e-44b0-bf0c-ec3ba2776c55\") " pod="metallb-system/speaker-zmg22" Feb 14 11:21:41 crc kubenswrapper[4904]: E0214 11:21:41.393127 4904 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Feb 14 11:21:41 crc kubenswrapper[4904]: E0214 11:21:41.393186 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-metrics-certs podName:90c810e2-d81e-44b0-bf0c-ec3ba2776c55 nodeName:}" failed. No retries permitted until 2026-02-14 11:21:41.893167427 +0000 UTC m=+692.705932088 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-metrics-certs") pod "speaker-zmg22" (UID: "90c810e2-d81e-44b0-bf0c-ec3ba2776c55") : secret "speaker-certs-secret" not found Feb 14 11:21:41 crc kubenswrapper[4904]: E0214 11:21:41.393494 4904 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 14 11:21:41 crc kubenswrapper[4904]: E0214 11:21:41.393536 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-memberlist podName:90c810e2-d81e-44b0-bf0c-ec3ba2776c55 nodeName:}" failed. No retries permitted until 2026-02-14 11:21:41.893526387 +0000 UTC m=+692.706291048 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-memberlist") pod "speaker-zmg22" (UID: "90c810e2-d81e-44b0-bf0c-ec3ba2776c55") : secret "metallb-memberlist" not found Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.394462 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-metallb-excludel2\") pod \"speaker-zmg22\" (UID: \"90c810e2-d81e-44b0-bf0c-ec3ba2776c55\") " pod="metallb-system/speaker-zmg22" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.397804 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ed02a813-b989-405f-87d8-6b4ed3539671-metrics-certs\") pod \"controller-69bbfbf88f-bf7zh\" (UID: \"ed02a813-b989-405f-87d8-6b4ed3539671\") " pod="metallb-system/controller-69bbfbf88f-bf7zh" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.399338 4904 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.413159 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ed02a813-b989-405f-87d8-6b4ed3539671-cert\") pod \"controller-69bbfbf88f-bf7zh\" (UID: \"ed02a813-b989-405f-87d8-6b4ed3539671\") " pod="metallb-system/controller-69bbfbf88f-bf7zh" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.417594 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls4gd\" (UniqueName: \"kubernetes.io/projected/ed02a813-b989-405f-87d8-6b4ed3539671-kube-api-access-ls4gd\") pod \"controller-69bbfbf88f-bf7zh\" (UID: \"ed02a813-b989-405f-87d8-6b4ed3539671\") " pod="metallb-system/controller-69bbfbf88f-bf7zh" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.430422 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5k76\" (UniqueName: \"kubernetes.io/projected/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-kube-api-access-r5k76\") pod \"speaker-zmg22\" (UID: \"90c810e2-d81e-44b0-bf0c-ec3ba2776c55\") " pod="metallb-system/speaker-zmg22" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.495864 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-bf7zh" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.800524 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b15d7113-f0f8-4033-a293-fc94960799c4-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-jzq7k\" (UID: \"b15d7113-f0f8-4033-a293-fc94960799c4\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jzq7k" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.806633 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b15d7113-f0f8-4033-a293-fc94960799c4-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-jzq7k\" (UID: \"b15d7113-f0f8-4033-a293-fc94960799c4\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jzq7k" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.889942 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-bf7zh"] Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.902223 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-metrics-certs\") pod \"speaker-zmg22\" (UID: \"90c810e2-d81e-44b0-bf0c-ec3ba2776c55\") " pod="metallb-system/speaker-zmg22" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.902291 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-memberlist\") pod \"speaker-zmg22\" (UID: \"90c810e2-d81e-44b0-bf0c-ec3ba2776c55\") " pod="metallb-system/speaker-zmg22" Feb 14 11:21:41 crc kubenswrapper[4904]: E0214 11:21:41.902411 4904 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 14 11:21:41 crc kubenswrapper[4904]: E0214 11:21:41.902485 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-memberlist podName:90c810e2-d81e-44b0-bf0c-ec3ba2776c55 nodeName:}" failed. No retries permitted until 2026-02-14 11:21:42.902463187 +0000 UTC m=+693.715227848 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-memberlist") pod "speaker-zmg22" (UID: "90c810e2-d81e-44b0-bf0c-ec3ba2776c55") : secret "metallb-memberlist" not found Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.906214 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-metrics-certs\") pod \"speaker-zmg22\" (UID: \"90c810e2-d81e-44b0-bf0c-ec3ba2776c55\") " pod="metallb-system/speaker-zmg22" Feb 14 11:21:41 crc kubenswrapper[4904]: I0214 11:21:41.963780 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jzq7k" Feb 14 11:21:42 crc kubenswrapper[4904]: I0214 11:21:42.113944 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6sv9" event={"ID":"bdf867d3-84cd-4b7c-8e49-97ba408a1689","Type":"ContainerStarted","Data":"b379f9a444b91fb2ec2667ac01dc7ebdbd8e2268cb5e5b7db747127e4b226748"} Feb 14 11:21:42 crc kubenswrapper[4904]: I0214 11:21:42.116826 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-bf7zh" event={"ID":"ed02a813-b989-405f-87d8-6b4ed3539671","Type":"ContainerStarted","Data":"7c2c126a834463c6e31037d1d2398b0172bd5f1dadece27d4fcdfe33c1a3aa0d"} Feb 14 11:21:42 crc kubenswrapper[4904]: I0214 11:21:42.116877 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-bf7zh" event={"ID":"ed02a813-b989-405f-87d8-6b4ed3539671","Type":"ContainerStarted","Data":"9b0e6f8e7b9306cd89ff0de313ba0a05cab8200eda6a5a470e0beeda688324a2"} Feb 14 11:21:42 crc kubenswrapper[4904]: I0214 11:21:42.231862 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-jzq7k"] Feb 14 11:21:42 crc kubenswrapper[4904]: W0214 11:21:42.236184 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb15d7113_f0f8_4033_a293_fc94960799c4.slice/crio-8fa470d9acdc673647705af861be5a493998299ad4aefe0f3f817111ec0bda9c WatchSource:0}: Error finding container 8fa470d9acdc673647705af861be5a493998299ad4aefe0f3f817111ec0bda9c: Status 404 returned error can't find the container with id 8fa470d9acdc673647705af861be5a493998299ad4aefe0f3f817111ec0bda9c Feb 14 11:21:42 crc kubenswrapper[4904]: I0214 11:21:42.919064 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-memberlist\") pod \"speaker-zmg22\" (UID: \"90c810e2-d81e-44b0-bf0c-ec3ba2776c55\") " pod="metallb-system/speaker-zmg22" Feb 14 11:21:42 crc kubenswrapper[4904]: I0214 11:21:42.923625 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/90c810e2-d81e-44b0-bf0c-ec3ba2776c55-memberlist\") pod \"speaker-zmg22\" (UID: \"90c810e2-d81e-44b0-bf0c-ec3ba2776c55\") " pod="metallb-system/speaker-zmg22" Feb 14 11:21:42 crc kubenswrapper[4904]: I0214 11:21:42.958366 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-zmg22" Feb 14 11:21:42 crc kubenswrapper[4904]: W0214 11:21:42.986156 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90c810e2_d81e_44b0_bf0c_ec3ba2776c55.slice/crio-5ee7e98690f854c3695eed480b723ad1845c4bc284275ace7b749d7634e90bd8 WatchSource:0}: Error finding container 5ee7e98690f854c3695eed480b723ad1845c4bc284275ace7b749d7634e90bd8: Status 404 returned error can't find the container with id 5ee7e98690f854c3695eed480b723ad1845c4bc284275ace7b749d7634e90bd8 Feb 14 11:21:43 crc kubenswrapper[4904]: I0214 11:21:43.142995 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jzq7k" event={"ID":"b15d7113-f0f8-4033-a293-fc94960799c4","Type":"ContainerStarted","Data":"8fa470d9acdc673647705af861be5a493998299ad4aefe0f3f817111ec0bda9c"} Feb 14 11:21:43 crc kubenswrapper[4904]: I0214 11:21:43.146775 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zmg22" event={"ID":"90c810e2-d81e-44b0-bf0c-ec3ba2776c55","Type":"ContainerStarted","Data":"5ee7e98690f854c3695eed480b723ad1845c4bc284275ace7b749d7634e90bd8"} Feb 14 11:21:43 crc kubenswrapper[4904]: I0214 11:21:43.149521 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-bf7zh" event={"ID":"ed02a813-b989-405f-87d8-6b4ed3539671","Type":"ContainerStarted","Data":"47079b72912e56b6eeb2a72ff121cb98e6bd1ed4cabb286c5999294f6d4e28b8"} Feb 14 11:21:43 crc kubenswrapper[4904]: I0214 11:21:43.150524 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-69bbfbf88f-bf7zh" Feb 14 11:21:43 crc kubenswrapper[4904]: I0214 11:21:43.170040 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-69bbfbf88f-bf7zh" podStartSLOduration=2.170027836 podStartE2EDuration="2.170027836s" podCreationTimestamp="2026-02-14 11:21:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:21:43.168614015 +0000 UTC m=+693.981378666" watchObservedRunningTime="2026-02-14 11:21:43.170027836 +0000 UTC m=+693.982792497" Feb 14 11:21:44 crc kubenswrapper[4904]: I0214 11:21:44.160105 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zmg22" event={"ID":"90c810e2-d81e-44b0-bf0c-ec3ba2776c55","Type":"ContainerStarted","Data":"e60147ded9205dd30a4008fbf2d78cd390c36a5ee8ea8a8c17b805f1ae28a2c0"} Feb 14 11:21:44 crc kubenswrapper[4904]: I0214 11:21:44.160376 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zmg22" event={"ID":"90c810e2-d81e-44b0-bf0c-ec3ba2776c55","Type":"ContainerStarted","Data":"036bc34698b4beccf5ff6df368f57026215f1744307480e758020aa7ba766c35"} Feb 14 11:21:44 crc kubenswrapper[4904]: I0214 11:21:44.160396 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-zmg22" Feb 14 11:21:44 crc kubenswrapper[4904]: I0214 11:21:44.177231 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-zmg22" podStartSLOduration=3.177211479 podStartE2EDuration="3.177211479s" podCreationTimestamp="2026-02-14 11:21:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:21:44.174103921 +0000 UTC m=+694.986868582" watchObservedRunningTime="2026-02-14 11:21:44.177211479 +0000 UTC m=+694.989976140" Feb 14 11:21:49 crc kubenswrapper[4904]: I0214 11:21:49.224447 4904 generic.go:334] "Generic (PLEG): container finished" podID="bdf867d3-84cd-4b7c-8e49-97ba408a1689" containerID="e8456fde0997ace867e9739a3b6795dc0c116124e0d1098a6e728f8748322168" exitCode=0 Feb 14 11:21:49 crc kubenswrapper[4904]: I0214 11:21:49.224703 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6sv9" event={"ID":"bdf867d3-84cd-4b7c-8e49-97ba408a1689","Type":"ContainerDied","Data":"e8456fde0997ace867e9739a3b6795dc0c116124e0d1098a6e728f8748322168"} Feb 14 11:21:49 crc kubenswrapper[4904]: I0214 11:21:49.227322 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jzq7k" event={"ID":"b15d7113-f0f8-4033-a293-fc94960799c4","Type":"ContainerStarted","Data":"73a038fa78f335f1feff8aea1f87110c5658bcf58f747509245f451ac0a8d220"} Feb 14 11:21:49 crc kubenswrapper[4904]: I0214 11:21:49.227949 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jzq7k" Feb 14 11:21:50 crc kubenswrapper[4904]: I0214 11:21:50.235407 4904 generic.go:334] "Generic (PLEG): container finished" podID="bdf867d3-84cd-4b7c-8e49-97ba408a1689" containerID="4c483380b0270b0f0b20c31c220159c5f73de4f21f128e1b859f6dca32383596" exitCode=0 Feb 14 11:21:50 crc kubenswrapper[4904]: I0214 11:21:50.235588 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6sv9" event={"ID":"bdf867d3-84cd-4b7c-8e49-97ba408a1689","Type":"ContainerDied","Data":"4c483380b0270b0f0b20c31c220159c5f73de4f21f128e1b859f6dca32383596"} Feb 14 11:21:50 crc kubenswrapper[4904]: I0214 11:21:50.257942 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jzq7k" podStartSLOduration=2.571163241 podStartE2EDuration="9.257921452s" podCreationTimestamp="2026-02-14 11:21:41 +0000 UTC" firstStartedPulling="2026-02-14 11:21:42.238284542 +0000 UTC m=+693.051049193" lastFinishedPulling="2026-02-14 11:21:48.925042613 +0000 UTC m=+699.737807404" observedRunningTime="2026-02-14 11:21:49.269197596 +0000 UTC m=+700.081962277" watchObservedRunningTime="2026-02-14 11:21:50.257921452 +0000 UTC m=+701.070686113" Feb 14 11:21:51 crc kubenswrapper[4904]: I0214 11:21:51.244827 4904 generic.go:334] "Generic (PLEG): container finished" podID="bdf867d3-84cd-4b7c-8e49-97ba408a1689" containerID="646748fcc16b80159e090c8b60c9ffb7d58a689307439882e1a60e7176de29d2" exitCode=0 Feb 14 11:21:51 crc kubenswrapper[4904]: I0214 11:21:51.244964 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6sv9" event={"ID":"bdf867d3-84cd-4b7c-8e49-97ba408a1689","Type":"ContainerDied","Data":"646748fcc16b80159e090c8b60c9ffb7d58a689307439882e1a60e7176de29d2"} Feb 14 11:21:52 crc kubenswrapper[4904]: I0214 11:21:52.256088 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6sv9" event={"ID":"bdf867d3-84cd-4b7c-8e49-97ba408a1689","Type":"ContainerStarted","Data":"26ab38d1ecf44591686c21548ed933299becc1bdc691d6d31c854cf40327ff82"} Feb 14 11:21:52 crc kubenswrapper[4904]: I0214 11:21:52.256127 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6sv9" event={"ID":"bdf867d3-84cd-4b7c-8e49-97ba408a1689","Type":"ContainerStarted","Data":"472e9469d080e9a6bf4315587b87c75424e3079fa7649ac365cdbec61707a3c5"} Feb 14 11:21:52 crc kubenswrapper[4904]: I0214 11:21:52.256138 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6sv9" event={"ID":"bdf867d3-84cd-4b7c-8e49-97ba408a1689","Type":"ContainerStarted","Data":"6943f2da3a3f3bacf2cfec4d47b79a642a2021fba7ca1488938e21ec8013a6a4"} Feb 14 11:21:52 crc kubenswrapper[4904]: I0214 11:21:52.256146 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6sv9" event={"ID":"bdf867d3-84cd-4b7c-8e49-97ba408a1689","Type":"ContainerStarted","Data":"66397c2a8fcee9f0be1399f5ed23ce62dfeadfd83d3f116dd10f98770c93eb0f"} Feb 14 11:21:52 crc kubenswrapper[4904]: I0214 11:21:52.256154 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6sv9" event={"ID":"bdf867d3-84cd-4b7c-8e49-97ba408a1689","Type":"ContainerStarted","Data":"8b9c04c788e0625b805aa86a84a10e8c41e9889f8757c4fa8978285401efc815"} Feb 14 11:21:52 crc kubenswrapper[4904]: I0214 11:21:52.256162 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-x6sv9" event={"ID":"bdf867d3-84cd-4b7c-8e49-97ba408a1689","Type":"ContainerStarted","Data":"2a5dc1f84c013649afbdf72e6ebe105b0f7a7063701a4371ca24fb3fea990fed"} Feb 14 11:21:52 crc kubenswrapper[4904]: I0214 11:21:52.256300 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:52 crc kubenswrapper[4904]: I0214 11:21:52.274680 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-x6sv9" podStartSLOduration=3.914370284 podStartE2EDuration="11.274660439s" podCreationTimestamp="2026-02-14 11:21:41 +0000 UTC" firstStartedPulling="2026-02-14 11:21:41.54210839 +0000 UTC m=+692.354873051" lastFinishedPulling="2026-02-14 11:21:48.902398545 +0000 UTC m=+699.715163206" observedRunningTime="2026-02-14 11:21:52.272150377 +0000 UTC m=+703.084915038" watchObservedRunningTime="2026-02-14 11:21:52.274660439 +0000 UTC m=+703.087425100" Feb 14 11:21:56 crc kubenswrapper[4904]: I0214 11:21:56.372999 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:21:56 crc kubenswrapper[4904]: I0214 11:21:56.430866 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:22:01 crc kubenswrapper[4904]: I0214 11:22:01.375605 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-x6sv9" Feb 14 11:22:01 crc kubenswrapper[4904]: I0214 11:22:01.502005 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-69bbfbf88f-bf7zh" Feb 14 11:22:01 crc kubenswrapper[4904]: I0214 11:22:01.968763 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jzq7k" Feb 14 11:22:02 crc kubenswrapper[4904]: I0214 11:22:02.961576 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-zmg22" Feb 14 11:22:05 crc kubenswrapper[4904]: I0214 11:22:05.850725 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-ngws6"] Feb 14 11:22:05 crc kubenswrapper[4904]: I0214 11:22:05.852060 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ngws6" Feb 14 11:22:05 crc kubenswrapper[4904]: I0214 11:22:05.855968 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-vvnmj" Feb 14 11:22:05 crc kubenswrapper[4904]: I0214 11:22:05.856582 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 14 11:22:05 crc kubenswrapper[4904]: I0214 11:22:05.856824 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 14 11:22:05 crc kubenswrapper[4904]: I0214 11:22:05.869412 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-ngws6"] Feb 14 11:22:05 crc kubenswrapper[4904]: I0214 11:22:05.996403 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn8f5\" (UniqueName: \"kubernetes.io/projected/b8af9363-bb1b-4bb6-9165-f1db445f7afc-kube-api-access-nn8f5\") pod \"openstack-operator-index-ngws6\" (UID: \"b8af9363-bb1b-4bb6-9165-f1db445f7afc\") " pod="openstack-operators/openstack-operator-index-ngws6" Feb 14 11:22:06 crc kubenswrapper[4904]: I0214 11:22:06.098829 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn8f5\" (UniqueName: \"kubernetes.io/projected/b8af9363-bb1b-4bb6-9165-f1db445f7afc-kube-api-access-nn8f5\") pod \"openstack-operator-index-ngws6\" (UID: \"b8af9363-bb1b-4bb6-9165-f1db445f7afc\") " pod="openstack-operators/openstack-operator-index-ngws6" Feb 14 11:22:06 crc kubenswrapper[4904]: I0214 11:22:06.122391 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn8f5\" (UniqueName: \"kubernetes.io/projected/b8af9363-bb1b-4bb6-9165-f1db445f7afc-kube-api-access-nn8f5\") pod \"openstack-operator-index-ngws6\" (UID: \"b8af9363-bb1b-4bb6-9165-f1db445f7afc\") " pod="openstack-operators/openstack-operator-index-ngws6" Feb 14 11:22:06 crc kubenswrapper[4904]: I0214 11:22:06.171237 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ngws6" Feb 14 11:22:06 crc kubenswrapper[4904]: I0214 11:22:06.583974 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-ngws6"] Feb 14 11:22:06 crc kubenswrapper[4904]: I0214 11:22:06.651984 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ngws6" event={"ID":"b8af9363-bb1b-4bb6-9165-f1db445f7afc","Type":"ContainerStarted","Data":"10e5eda52b5715ab6268feb714c555104856e0eb244303f147c71e518c5ff9ec"} Feb 14 11:22:09 crc kubenswrapper[4904]: I0214 11:22:09.023367 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-ngws6"] Feb 14 11:22:09 crc kubenswrapper[4904]: I0214 11:22:09.630185 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-r9h6q"] Feb 14 11:22:09 crc kubenswrapper[4904]: I0214 11:22:09.631356 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r9h6q" Feb 14 11:22:09 crc kubenswrapper[4904]: I0214 11:22:09.655154 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-r9h6q"] Feb 14 11:22:09 crc kubenswrapper[4904]: I0214 11:22:09.680259 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ngws6" event={"ID":"b8af9363-bb1b-4bb6-9165-f1db445f7afc","Type":"ContainerStarted","Data":"2a2f6e0ebd81a67a85741a4e6666b8227e19e2067756019ecce900144d295666"} Feb 14 11:22:09 crc kubenswrapper[4904]: I0214 11:22:09.680387 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-ngws6" podUID="b8af9363-bb1b-4bb6-9165-f1db445f7afc" containerName="registry-server" containerID="cri-o://2a2f6e0ebd81a67a85741a4e6666b8227e19e2067756019ecce900144d295666" gracePeriod=2 Feb 14 11:22:09 crc kubenswrapper[4904]: I0214 11:22:09.698378 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-ngws6" podStartSLOduration=2.34786341 podStartE2EDuration="4.698362526s" podCreationTimestamp="2026-02-14 11:22:05 +0000 UTC" firstStartedPulling="2026-02-14 11:22:06.595952831 +0000 UTC m=+717.408717482" lastFinishedPulling="2026-02-14 11:22:08.946451937 +0000 UTC m=+719.759216598" observedRunningTime="2026-02-14 11:22:09.695171998 +0000 UTC m=+720.507936659" watchObservedRunningTime="2026-02-14 11:22:09.698362526 +0000 UTC m=+720.511127187" Feb 14 11:22:09 crc kubenswrapper[4904]: I0214 11:22:09.769421 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7qrq\" (UniqueName: \"kubernetes.io/projected/b429a7c8-df33-430d-9d4f-a678e2639575-kube-api-access-t7qrq\") pod \"openstack-operator-index-r9h6q\" (UID: \"b429a7c8-df33-430d-9d4f-a678e2639575\") " pod="openstack-operators/openstack-operator-index-r9h6q" Feb 14 11:22:09 crc kubenswrapper[4904]: I0214 11:22:09.872455 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7qrq\" (UniqueName: \"kubernetes.io/projected/b429a7c8-df33-430d-9d4f-a678e2639575-kube-api-access-t7qrq\") pod \"openstack-operator-index-r9h6q\" (UID: \"b429a7c8-df33-430d-9d4f-a678e2639575\") " pod="openstack-operators/openstack-operator-index-r9h6q" Feb 14 11:22:09 crc kubenswrapper[4904]: I0214 11:22:09.898483 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7qrq\" (UniqueName: \"kubernetes.io/projected/b429a7c8-df33-430d-9d4f-a678e2639575-kube-api-access-t7qrq\") pod \"openstack-operator-index-r9h6q\" (UID: \"b429a7c8-df33-430d-9d4f-a678e2639575\") " pod="openstack-operators/openstack-operator-index-r9h6q" Feb 14 11:22:09 crc kubenswrapper[4904]: I0214 11:22:09.950187 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r9h6q" Feb 14 11:22:10 crc kubenswrapper[4904]: I0214 11:22:10.077456 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ngws6" Feb 14 11:22:10 crc kubenswrapper[4904]: I0214 11:22:10.175310 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nn8f5\" (UniqueName: \"kubernetes.io/projected/b8af9363-bb1b-4bb6-9165-f1db445f7afc-kube-api-access-nn8f5\") pod \"b8af9363-bb1b-4bb6-9165-f1db445f7afc\" (UID: \"b8af9363-bb1b-4bb6-9165-f1db445f7afc\") " Feb 14 11:22:10 crc kubenswrapper[4904]: I0214 11:22:10.178495 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8af9363-bb1b-4bb6-9165-f1db445f7afc-kube-api-access-nn8f5" (OuterVolumeSpecName: "kube-api-access-nn8f5") pod "b8af9363-bb1b-4bb6-9165-f1db445f7afc" (UID: "b8af9363-bb1b-4bb6-9165-f1db445f7afc"). InnerVolumeSpecName "kube-api-access-nn8f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:22:10 crc kubenswrapper[4904]: I0214 11:22:10.276490 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nn8f5\" (UniqueName: \"kubernetes.io/projected/b8af9363-bb1b-4bb6-9165-f1db445f7afc-kube-api-access-nn8f5\") on node \"crc\" DevicePath \"\"" Feb 14 11:22:10 crc kubenswrapper[4904]: I0214 11:22:10.433880 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-r9h6q"] Feb 14 11:22:10 crc kubenswrapper[4904]: W0214 11:22:10.435531 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb429a7c8_df33_430d_9d4f_a678e2639575.slice/crio-dd736cf806528955841686cb6255daaee37fc36cdd27eed04133745b8e1100bf WatchSource:0}: Error finding container dd736cf806528955841686cb6255daaee37fc36cdd27eed04133745b8e1100bf: Status 404 returned error can't find the container with id dd736cf806528955841686cb6255daaee37fc36cdd27eed04133745b8e1100bf Feb 14 11:22:10 crc kubenswrapper[4904]: I0214 11:22:10.689015 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r9h6q" event={"ID":"b429a7c8-df33-430d-9d4f-a678e2639575","Type":"ContainerStarted","Data":"82fecb48526f651745c962a62f809076f73c5bcfc63633bb6718e94c2008414f"} Feb 14 11:22:10 crc kubenswrapper[4904]: I0214 11:22:10.689057 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r9h6q" event={"ID":"b429a7c8-df33-430d-9d4f-a678e2639575","Type":"ContainerStarted","Data":"dd736cf806528955841686cb6255daaee37fc36cdd27eed04133745b8e1100bf"} Feb 14 11:22:10 crc kubenswrapper[4904]: I0214 11:22:10.690662 4904 generic.go:334] "Generic (PLEG): container finished" podID="b8af9363-bb1b-4bb6-9165-f1db445f7afc" containerID="2a2f6e0ebd81a67a85741a4e6666b8227e19e2067756019ecce900144d295666" exitCode=0 Feb 14 11:22:10 crc kubenswrapper[4904]: I0214 11:22:10.690709 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ngws6" event={"ID":"b8af9363-bb1b-4bb6-9165-f1db445f7afc","Type":"ContainerDied","Data":"2a2f6e0ebd81a67a85741a4e6666b8227e19e2067756019ecce900144d295666"} Feb 14 11:22:10 crc kubenswrapper[4904]: I0214 11:22:10.690734 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ngws6" event={"ID":"b8af9363-bb1b-4bb6-9165-f1db445f7afc","Type":"ContainerDied","Data":"10e5eda52b5715ab6268feb714c555104856e0eb244303f147c71e518c5ff9ec"} Feb 14 11:22:10 crc kubenswrapper[4904]: I0214 11:22:10.690750 4904 scope.go:117] "RemoveContainer" containerID="2a2f6e0ebd81a67a85741a4e6666b8227e19e2067756019ecce900144d295666" Feb 14 11:22:10 crc kubenswrapper[4904]: I0214 11:22:10.691071 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ngws6" Feb 14 11:22:10 crc kubenswrapper[4904]: I0214 11:22:10.708209 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-r9h6q" podStartSLOduration=1.6655417620000001 podStartE2EDuration="1.708192174s" podCreationTimestamp="2026-02-14 11:22:09 +0000 UTC" firstStartedPulling="2026-02-14 11:22:10.440150635 +0000 UTC m=+721.252915296" lastFinishedPulling="2026-02-14 11:22:10.482801047 +0000 UTC m=+721.295565708" observedRunningTime="2026-02-14 11:22:10.703925256 +0000 UTC m=+721.516689927" watchObservedRunningTime="2026-02-14 11:22:10.708192174 +0000 UTC m=+721.520956835" Feb 14 11:22:10 crc kubenswrapper[4904]: I0214 11:22:10.708801 4904 scope.go:117] "RemoveContainer" containerID="2a2f6e0ebd81a67a85741a4e6666b8227e19e2067756019ecce900144d295666" Feb 14 11:22:10 crc kubenswrapper[4904]: E0214 11:22:10.709298 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a2f6e0ebd81a67a85741a4e6666b8227e19e2067756019ecce900144d295666\": container with ID starting with 2a2f6e0ebd81a67a85741a4e6666b8227e19e2067756019ecce900144d295666 not found: ID does not exist" containerID="2a2f6e0ebd81a67a85741a4e6666b8227e19e2067756019ecce900144d295666" Feb 14 11:22:10 crc kubenswrapper[4904]: I0214 11:22:10.709367 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a2f6e0ebd81a67a85741a4e6666b8227e19e2067756019ecce900144d295666"} err="failed to get container status \"2a2f6e0ebd81a67a85741a4e6666b8227e19e2067756019ecce900144d295666\": rpc error: code = NotFound desc = could not find container \"2a2f6e0ebd81a67a85741a4e6666b8227e19e2067756019ecce900144d295666\": container with ID starting with 2a2f6e0ebd81a67a85741a4e6666b8227e19e2067756019ecce900144d295666 not found: ID does not exist" Feb 14 11:22:10 crc kubenswrapper[4904]: I0214 11:22:10.725946 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-ngws6"] Feb 14 11:22:10 crc kubenswrapper[4904]: I0214 11:22:10.730648 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-ngws6"] Feb 14 11:22:11 crc kubenswrapper[4904]: I0214 11:22:11.845088 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8af9363-bb1b-4bb6-9165-f1db445f7afc" path="/var/lib/kubelet/pods/b8af9363-bb1b-4bb6-9165-f1db445f7afc/volumes" Feb 14 11:22:16 crc kubenswrapper[4904]: I0214 11:22:16.382510 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:22:16 crc kubenswrapper[4904]: I0214 11:22:16.382800 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:22:19 crc kubenswrapper[4904]: I0214 11:22:19.951148 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-r9h6q" Feb 14 11:22:19 crc kubenswrapper[4904]: I0214 11:22:19.952014 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-r9h6q" Feb 14 11:22:19 crc kubenswrapper[4904]: I0214 11:22:19.975936 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-r9h6q" Feb 14 11:22:20 crc kubenswrapper[4904]: I0214 11:22:20.783287 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-r9h6q" Feb 14 11:22:27 crc kubenswrapper[4904]: I0214 11:22:27.433643 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn"] Feb 14 11:22:27 crc kubenswrapper[4904]: E0214 11:22:27.434396 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8af9363-bb1b-4bb6-9165-f1db445f7afc" containerName="registry-server" Feb 14 11:22:27 crc kubenswrapper[4904]: I0214 11:22:27.434409 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8af9363-bb1b-4bb6-9165-f1db445f7afc" containerName="registry-server" Feb 14 11:22:27 crc kubenswrapper[4904]: I0214 11:22:27.434560 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8af9363-bb1b-4bb6-9165-f1db445f7afc" containerName="registry-server" Feb 14 11:22:27 crc kubenswrapper[4904]: I0214 11:22:27.435421 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn" Feb 14 11:22:27 crc kubenswrapper[4904]: I0214 11:22:27.448270 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-62kx6" Feb 14 11:22:27 crc kubenswrapper[4904]: I0214 11:22:27.450138 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn"] Feb 14 11:22:27 crc kubenswrapper[4904]: I0214 11:22:27.526160 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gcsf\" (UniqueName: \"kubernetes.io/projected/a0bea6e6-38b1-427d-8699-a29ff81030a7-kube-api-access-2gcsf\") pod \"e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn\" (UID: \"a0bea6e6-38b1-427d-8699-a29ff81030a7\") " pod="openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn" Feb 14 11:22:27 crc kubenswrapper[4904]: I0214 11:22:27.526300 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a0bea6e6-38b1-427d-8699-a29ff81030a7-bundle\") pod \"e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn\" (UID: \"a0bea6e6-38b1-427d-8699-a29ff81030a7\") " pod="openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn" Feb 14 11:22:27 crc kubenswrapper[4904]: I0214 11:22:27.526378 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a0bea6e6-38b1-427d-8699-a29ff81030a7-util\") pod \"e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn\" (UID: \"a0bea6e6-38b1-427d-8699-a29ff81030a7\") " pod="openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn" Feb 14 11:22:27 crc kubenswrapper[4904]: I0214 11:22:27.627393 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gcsf\" (UniqueName: \"kubernetes.io/projected/a0bea6e6-38b1-427d-8699-a29ff81030a7-kube-api-access-2gcsf\") pod \"e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn\" (UID: \"a0bea6e6-38b1-427d-8699-a29ff81030a7\") " pod="openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn" Feb 14 11:22:27 crc kubenswrapper[4904]: I0214 11:22:27.627448 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a0bea6e6-38b1-427d-8699-a29ff81030a7-bundle\") pod \"e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn\" (UID: \"a0bea6e6-38b1-427d-8699-a29ff81030a7\") " pod="openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn" Feb 14 11:22:27 crc kubenswrapper[4904]: I0214 11:22:27.627497 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a0bea6e6-38b1-427d-8699-a29ff81030a7-util\") pod \"e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn\" (UID: \"a0bea6e6-38b1-427d-8699-a29ff81030a7\") " pod="openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn" Feb 14 11:22:27 crc kubenswrapper[4904]: I0214 11:22:27.627898 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a0bea6e6-38b1-427d-8699-a29ff81030a7-bundle\") pod \"e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn\" (UID: \"a0bea6e6-38b1-427d-8699-a29ff81030a7\") " pod="openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn" Feb 14 11:22:27 crc kubenswrapper[4904]: I0214 11:22:27.627947 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a0bea6e6-38b1-427d-8699-a29ff81030a7-util\") pod \"e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn\" (UID: \"a0bea6e6-38b1-427d-8699-a29ff81030a7\") " pod="openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn" Feb 14 11:22:27 crc kubenswrapper[4904]: I0214 11:22:27.646648 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gcsf\" (UniqueName: \"kubernetes.io/projected/a0bea6e6-38b1-427d-8699-a29ff81030a7-kube-api-access-2gcsf\") pod \"e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn\" (UID: \"a0bea6e6-38b1-427d-8699-a29ff81030a7\") " pod="openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn" Feb 14 11:22:27 crc kubenswrapper[4904]: I0214 11:22:27.762641 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn" Feb 14 11:22:28 crc kubenswrapper[4904]: I0214 11:22:28.157385 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn"] Feb 14 11:22:28 crc kubenswrapper[4904]: I0214 11:22:28.800819 4904 generic.go:334] "Generic (PLEG): container finished" podID="a0bea6e6-38b1-427d-8699-a29ff81030a7" containerID="b81a2737f6a3953b2364d3eb97282819ab1ea495cc5010f2dfdda6d05cfa6932" exitCode=0 Feb 14 11:22:28 crc kubenswrapper[4904]: I0214 11:22:28.800907 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn" event={"ID":"a0bea6e6-38b1-427d-8699-a29ff81030a7","Type":"ContainerDied","Data":"b81a2737f6a3953b2364d3eb97282819ab1ea495cc5010f2dfdda6d05cfa6932"} Feb 14 11:22:28 crc kubenswrapper[4904]: I0214 11:22:28.801245 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn" event={"ID":"a0bea6e6-38b1-427d-8699-a29ff81030a7","Type":"ContainerStarted","Data":"e8e5cb6e748a649e5a90ec5ba5343b0b0623ccbbbed1c73a1674578002657a6b"} Feb 14 11:22:29 crc kubenswrapper[4904]: I0214 11:22:29.812711 4904 generic.go:334] "Generic (PLEG): container finished" podID="a0bea6e6-38b1-427d-8699-a29ff81030a7" containerID="c727ba28379b2dc3458b459fe678f822252206889e568b3c8f1deefd5e46724f" exitCode=0 Feb 14 11:22:29 crc kubenswrapper[4904]: I0214 11:22:29.812807 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn" event={"ID":"a0bea6e6-38b1-427d-8699-a29ff81030a7","Type":"ContainerDied","Data":"c727ba28379b2dc3458b459fe678f822252206889e568b3c8f1deefd5e46724f"} Feb 14 11:22:30 crc kubenswrapper[4904]: I0214 11:22:30.819851 4904 generic.go:334] "Generic (PLEG): container finished" podID="a0bea6e6-38b1-427d-8699-a29ff81030a7" containerID="7871bfe5b35ba397df5f23c6f9d6b6e5f8fd9a74e002e7bbe2fde4150bfc8446" exitCode=0 Feb 14 11:22:30 crc kubenswrapper[4904]: I0214 11:22:30.819941 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn" event={"ID":"a0bea6e6-38b1-427d-8699-a29ff81030a7","Type":"ContainerDied","Data":"7871bfe5b35ba397df5f23c6f9d6b6e5f8fd9a74e002e7bbe2fde4150bfc8446"} Feb 14 11:22:32 crc kubenswrapper[4904]: I0214 11:22:32.098184 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn" Feb 14 11:22:32 crc kubenswrapper[4904]: I0214 11:22:32.190021 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gcsf\" (UniqueName: \"kubernetes.io/projected/a0bea6e6-38b1-427d-8699-a29ff81030a7-kube-api-access-2gcsf\") pod \"a0bea6e6-38b1-427d-8699-a29ff81030a7\" (UID: \"a0bea6e6-38b1-427d-8699-a29ff81030a7\") " Feb 14 11:22:32 crc kubenswrapper[4904]: I0214 11:22:32.190075 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a0bea6e6-38b1-427d-8699-a29ff81030a7-bundle\") pod \"a0bea6e6-38b1-427d-8699-a29ff81030a7\" (UID: \"a0bea6e6-38b1-427d-8699-a29ff81030a7\") " Feb 14 11:22:32 crc kubenswrapper[4904]: I0214 11:22:32.190112 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a0bea6e6-38b1-427d-8699-a29ff81030a7-util\") pod \"a0bea6e6-38b1-427d-8699-a29ff81030a7\" (UID: \"a0bea6e6-38b1-427d-8699-a29ff81030a7\") " Feb 14 11:22:32 crc kubenswrapper[4904]: I0214 11:22:32.191063 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0bea6e6-38b1-427d-8699-a29ff81030a7-bundle" (OuterVolumeSpecName: "bundle") pod "a0bea6e6-38b1-427d-8699-a29ff81030a7" (UID: "a0bea6e6-38b1-427d-8699-a29ff81030a7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:22:32 crc kubenswrapper[4904]: I0214 11:22:32.196979 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0bea6e6-38b1-427d-8699-a29ff81030a7-kube-api-access-2gcsf" (OuterVolumeSpecName: "kube-api-access-2gcsf") pod "a0bea6e6-38b1-427d-8699-a29ff81030a7" (UID: "a0bea6e6-38b1-427d-8699-a29ff81030a7"). InnerVolumeSpecName "kube-api-access-2gcsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:22:32 crc kubenswrapper[4904]: I0214 11:22:32.203781 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0bea6e6-38b1-427d-8699-a29ff81030a7-util" (OuterVolumeSpecName: "util") pod "a0bea6e6-38b1-427d-8699-a29ff81030a7" (UID: "a0bea6e6-38b1-427d-8699-a29ff81030a7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:22:32 crc kubenswrapper[4904]: I0214 11:22:32.291143 4904 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a0bea6e6-38b1-427d-8699-a29ff81030a7-util\") on node \"crc\" DevicePath \"\"" Feb 14 11:22:32 crc kubenswrapper[4904]: I0214 11:22:32.291177 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gcsf\" (UniqueName: \"kubernetes.io/projected/a0bea6e6-38b1-427d-8699-a29ff81030a7-kube-api-access-2gcsf\") on node \"crc\" DevicePath \"\"" Feb 14 11:22:32 crc kubenswrapper[4904]: I0214 11:22:32.291188 4904 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a0bea6e6-38b1-427d-8699-a29ff81030a7-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:22:32 crc kubenswrapper[4904]: I0214 11:22:32.839664 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn" event={"ID":"a0bea6e6-38b1-427d-8699-a29ff81030a7","Type":"ContainerDied","Data":"e8e5cb6e748a649e5a90ec5ba5343b0b0623ccbbbed1c73a1674578002657a6b"} Feb 14 11:22:32 crc kubenswrapper[4904]: I0214 11:22:32.839701 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8e5cb6e748a649e5a90ec5ba5343b0b0623ccbbbed1c73a1674578002657a6b" Feb 14 11:22:32 crc kubenswrapper[4904]: I0214 11:22:32.839755 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn" Feb 14 11:22:39 crc kubenswrapper[4904]: I0214 11:22:39.521658 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-bfd897969-pqr4v"] Feb 14 11:22:39 crc kubenswrapper[4904]: E0214 11:22:39.522238 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0bea6e6-38b1-427d-8699-a29ff81030a7" containerName="pull" Feb 14 11:22:39 crc kubenswrapper[4904]: I0214 11:22:39.522254 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0bea6e6-38b1-427d-8699-a29ff81030a7" containerName="pull" Feb 14 11:22:39 crc kubenswrapper[4904]: E0214 11:22:39.522272 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0bea6e6-38b1-427d-8699-a29ff81030a7" containerName="extract" Feb 14 11:22:39 crc kubenswrapper[4904]: I0214 11:22:39.522280 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0bea6e6-38b1-427d-8699-a29ff81030a7" containerName="extract" Feb 14 11:22:39 crc kubenswrapper[4904]: E0214 11:22:39.522293 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0bea6e6-38b1-427d-8699-a29ff81030a7" containerName="util" Feb 14 11:22:39 crc kubenswrapper[4904]: I0214 11:22:39.522301 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0bea6e6-38b1-427d-8699-a29ff81030a7" containerName="util" Feb 14 11:22:39 crc kubenswrapper[4904]: I0214 11:22:39.522432 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0bea6e6-38b1-427d-8699-a29ff81030a7" containerName="extract" Feb 14 11:22:39 crc kubenswrapper[4904]: I0214 11:22:39.522966 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-bfd897969-pqr4v" Feb 14 11:22:39 crc kubenswrapper[4904]: I0214 11:22:39.524935 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-djwb2" Feb 14 11:22:39 crc kubenswrapper[4904]: I0214 11:22:39.591512 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-bfd897969-pqr4v"] Feb 14 11:22:39 crc kubenswrapper[4904]: I0214 11:22:39.596065 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l76m\" (UniqueName: \"kubernetes.io/projected/6544b724-ca9a-4dec-9d1e-987e8b9a375b-kube-api-access-4l76m\") pod \"openstack-operator-controller-init-bfd897969-pqr4v\" (UID: \"6544b724-ca9a-4dec-9d1e-987e8b9a375b\") " pod="openstack-operators/openstack-operator-controller-init-bfd897969-pqr4v" Feb 14 11:22:39 crc kubenswrapper[4904]: I0214 11:22:39.696986 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l76m\" (UniqueName: \"kubernetes.io/projected/6544b724-ca9a-4dec-9d1e-987e8b9a375b-kube-api-access-4l76m\") pod \"openstack-operator-controller-init-bfd897969-pqr4v\" (UID: \"6544b724-ca9a-4dec-9d1e-987e8b9a375b\") " pod="openstack-operators/openstack-operator-controller-init-bfd897969-pqr4v" Feb 14 11:22:39 crc kubenswrapper[4904]: I0214 11:22:39.723982 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l76m\" (UniqueName: \"kubernetes.io/projected/6544b724-ca9a-4dec-9d1e-987e8b9a375b-kube-api-access-4l76m\") pod \"openstack-operator-controller-init-bfd897969-pqr4v\" (UID: \"6544b724-ca9a-4dec-9d1e-987e8b9a375b\") " pod="openstack-operators/openstack-operator-controller-init-bfd897969-pqr4v" Feb 14 11:22:39 crc kubenswrapper[4904]: I0214 11:22:39.837436 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-bfd897969-pqr4v" Feb 14 11:22:40 crc kubenswrapper[4904]: I0214 11:22:40.109395 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-bfd897969-pqr4v"] Feb 14 11:22:40 crc kubenswrapper[4904]: W0214 11:22:40.120765 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6544b724_ca9a_4dec_9d1e_987e8b9a375b.slice/crio-c7f15faf35f381211570ea0bc22b76fa47c59d2f57195d47c75c7aef6338a777 WatchSource:0}: Error finding container c7f15faf35f381211570ea0bc22b76fa47c59d2f57195d47c75c7aef6338a777: Status 404 returned error can't find the container with id c7f15faf35f381211570ea0bc22b76fa47c59d2f57195d47c75c7aef6338a777 Feb 14 11:22:40 crc kubenswrapper[4904]: I0214 11:22:40.905445 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-bfd897969-pqr4v" event={"ID":"6544b724-ca9a-4dec-9d1e-987e8b9a375b","Type":"ContainerStarted","Data":"c7f15faf35f381211570ea0bc22b76fa47c59d2f57195d47c75c7aef6338a777"} Feb 14 11:22:44 crc kubenswrapper[4904]: I0214 11:22:44.025306 4904 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 14 11:22:44 crc kubenswrapper[4904]: I0214 11:22:44.930161 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-bfd897969-pqr4v" event={"ID":"6544b724-ca9a-4dec-9d1e-987e8b9a375b","Type":"ContainerStarted","Data":"6c6757ddfca8ce88d6a1e696afd8453efc4900661614f1b97b8620120651c564"} Feb 14 11:22:44 crc kubenswrapper[4904]: I0214 11:22:44.930556 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-bfd897969-pqr4v" Feb 14 11:22:44 crc kubenswrapper[4904]: I0214 11:22:44.965776 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-bfd897969-pqr4v" podStartSLOduration=1.612038901 podStartE2EDuration="5.965752827s" podCreationTimestamp="2026-02-14 11:22:39 +0000 UTC" firstStartedPulling="2026-02-14 11:22:40.122558325 +0000 UTC m=+750.935322986" lastFinishedPulling="2026-02-14 11:22:44.476272261 +0000 UTC m=+755.289036912" observedRunningTime="2026-02-14 11:22:44.959110952 +0000 UTC m=+755.771875613" watchObservedRunningTime="2026-02-14 11:22:44.965752827 +0000 UTC m=+755.778517508" Feb 14 11:22:46 crc kubenswrapper[4904]: I0214 11:22:46.382948 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:22:46 crc kubenswrapper[4904]: I0214 11:22:46.383013 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:22:49 crc kubenswrapper[4904]: I0214 11:22:49.848012 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-bfd897969-pqr4v" Feb 14 11:23:16 crc kubenswrapper[4904]: I0214 11:23:16.382341 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:23:16 crc kubenswrapper[4904]: I0214 11:23:16.383724 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:23:16 crc kubenswrapper[4904]: I0214 11:23:16.384205 4904 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:23:16 crc kubenswrapper[4904]: I0214 11:23:16.384922 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ed0b33120210f8d03c6d331d3186cbee96523120ba158013d88a03648716b22f"} pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 14 11:23:16 crc kubenswrapper[4904]: I0214 11:23:16.385076 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" containerID="cri-o://ed0b33120210f8d03c6d331d3186cbee96523120ba158013d88a03648716b22f" gracePeriod=600 Feb 14 11:23:17 crc kubenswrapper[4904]: I0214 11:23:17.150385 4904 generic.go:334] "Generic (PLEG): container finished" podID="7790d276-2506-4641-8c22-67e6441417f0" containerID="ed0b33120210f8d03c6d331d3186cbee96523120ba158013d88a03648716b22f" exitCode=0 Feb 14 11:23:17 crc kubenswrapper[4904]: I0214 11:23:17.150467 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerDied","Data":"ed0b33120210f8d03c6d331d3186cbee96523120ba158013d88a03648716b22f"} Feb 14 11:23:17 crc kubenswrapper[4904]: I0214 11:23:17.150857 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"fd7271c1e14d135193bc38a4ef3eb1646cd74f3540f5fa874f75aef55d03613d"} Feb 14 11:23:17 crc kubenswrapper[4904]: I0214 11:23:17.150880 4904 scope.go:117] "RemoveContainer" containerID="7bee02c84077b982480197f3573826a9bf5c596ca426336719e5c06a6c3588f2" Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.763619 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-m29fl"] Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.765019 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-m29fl" Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.772783 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-2qhlk" Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.774929 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-hbqp9"] Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.775879 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-hbqp9" Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.779076 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-dlxks" Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.784580 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-gt547"] Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.785386 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-gt547" Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.789052 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-hf5sf" Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.801609 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-gt547"] Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.811201 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-hbqp9"] Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.816517 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-jsn2d"] Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.817432 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987464f4-jsn2d" Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.819424 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-7f82k" Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.914794 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-m29fl"] Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.927849 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7c4r\" (UniqueName: \"kubernetes.io/projected/33670c4d-fd2a-49cd-b602-ccdd927b7c00-kube-api-access-x7c4r\") pod \"cinder-operator-controller-manager-5d946d989d-hbqp9\" (UID: \"33670c4d-fd2a-49cd-b602-ccdd927b7c00\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-hbqp9" Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.927890 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clvtw\" (UniqueName: \"kubernetes.io/projected/6c57743e-c864-4e0d-828a-406978aab771-kube-api-access-clvtw\") pod \"designate-operator-controller-manager-6d8bf5c495-gt547\" (UID: \"6c57743e-c864-4e0d-828a-406978aab771\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-gt547" Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.927910 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9twzr\" (UniqueName: \"kubernetes.io/projected/2db31100-df1f-441c-87bf-944dc247a4fc-kube-api-access-9twzr\") pod \"glance-operator-controller-manager-77987464f4-jsn2d\" (UID: \"2db31100-df1f-441c-87bf-944dc247a4fc\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-jsn2d" Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.927957 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57zlf\" (UniqueName: \"kubernetes.io/projected/2b26d95c-682d-4be5-9788-5044e186611c-kube-api-access-57zlf\") pod \"barbican-operator-controller-manager-868647ff47-m29fl\" (UID: \"2b26d95c-682d-4be5-9788-5044e186611c\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-m29fl" Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.963726 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-mbrvg"] Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.964625 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-mbrvg" Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.968240 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-4shjl" Feb 14 11:23:25 crc kubenswrapper[4904]: I0214 11:23:25.973160 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-jsn2d"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.002079 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-m8p4r"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.003236 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-m8p4r" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.010167 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-nz42g" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.026936 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-mbrvg"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.028724 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7c4r\" (UniqueName: \"kubernetes.io/projected/33670c4d-fd2a-49cd-b602-ccdd927b7c00-kube-api-access-x7c4r\") pod \"cinder-operator-controller-manager-5d946d989d-hbqp9\" (UID: \"33670c4d-fd2a-49cd-b602-ccdd927b7c00\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-hbqp9" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.028858 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clvtw\" (UniqueName: \"kubernetes.io/projected/6c57743e-c864-4e0d-828a-406978aab771-kube-api-access-clvtw\") pod \"designate-operator-controller-manager-6d8bf5c495-gt547\" (UID: \"6c57743e-c864-4e0d-828a-406978aab771\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-gt547" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.028950 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9twzr\" (UniqueName: \"kubernetes.io/projected/2db31100-df1f-441c-87bf-944dc247a4fc-kube-api-access-9twzr\") pod \"glance-operator-controller-manager-77987464f4-jsn2d\" (UID: \"2db31100-df1f-441c-87bf-944dc247a4fc\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-jsn2d" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.029066 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57zlf\" (UniqueName: \"kubernetes.io/projected/2b26d95c-682d-4be5-9788-5044e186611c-kube-api-access-57zlf\") pod \"barbican-operator-controller-manager-868647ff47-m29fl\" (UID: \"2b26d95c-682d-4be5-9788-5044e186611c\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-m29fl" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.070277 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57zlf\" (UniqueName: \"kubernetes.io/projected/2b26d95c-682d-4be5-9788-5044e186611c-kube-api-access-57zlf\") pod \"barbican-operator-controller-manager-868647ff47-m29fl\" (UID: \"2b26d95c-682d-4be5-9788-5044e186611c\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-m29fl" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.070341 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-m8p4r"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.076583 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7c4r\" (UniqueName: \"kubernetes.io/projected/33670c4d-fd2a-49cd-b602-ccdd927b7c00-kube-api-access-x7c4r\") pod \"cinder-operator-controller-manager-5d946d989d-hbqp9\" (UID: \"33670c4d-fd2a-49cd-b602-ccdd927b7c00\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-hbqp9" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.082588 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-m29fl" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.086419 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.087326 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.090929 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.091366 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-s5cmc" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.093035 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clvtw\" (UniqueName: \"kubernetes.io/projected/6c57743e-c864-4e0d-828a-406978aab771-kube-api-access-clvtw\") pod \"designate-operator-controller-manager-6d8bf5c495-gt547\" (UID: \"6c57743e-c864-4e0d-828a-406978aab771\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-gt547" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.096716 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-k28jp"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.099486 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-hbqp9" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.100581 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.100656 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-k28jp" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.106194 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-gt547" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.131101 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbjqp\" (UniqueName: \"kubernetes.io/projected/d4632ec1-624e-4097-9888-0bbe395d6916-kube-api-access-sbjqp\") pod \"horizon-operator-controller-manager-5b9b8895d5-m8p4r\" (UID: \"d4632ec1-624e-4097-9888-0bbe395d6916\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-m8p4r" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.131215 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4xjq\" (UniqueName: \"kubernetes.io/projected/96c0bf26-aa73-4948-add8-b30fbbc23df9-kube-api-access-s4xjq\") pod \"heat-operator-controller-manager-69f49c598c-mbrvg\" (UID: \"96c0bf26-aa73-4948-add8-b30fbbc23df9\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-mbrvg" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.131300 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-mwspq" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.132905 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9twzr\" (UniqueName: \"kubernetes.io/projected/2db31100-df1f-441c-87bf-944dc247a4fc-kube-api-access-9twzr\") pod \"glance-operator-controller-manager-77987464f4-jsn2d\" (UID: \"2db31100-df1f-441c-87bf-944dc247a4fc\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-jsn2d" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.136474 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-k28jp"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.187898 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-sjlh5"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.188687 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-sjlh5" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.191504 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-q8zbv" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.202179 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-sjlh5"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.218913 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-6hjt5"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.220012 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-6hjt5" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.231031 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-rkjp2" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.232093 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987464f4-jsn2d" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.233000 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-cert\") pod \"infra-operator-controller-manager-5478cb7c68-h92m8\" (UID: \"8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9\") " pod="openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.233059 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4xjq\" (UniqueName: \"kubernetes.io/projected/96c0bf26-aa73-4948-add8-b30fbbc23df9-kube-api-access-s4xjq\") pod \"heat-operator-controller-manager-69f49c598c-mbrvg\" (UID: \"96c0bf26-aa73-4948-add8-b30fbbc23df9\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-mbrvg" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.233096 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw2n6\" (UniqueName: \"kubernetes.io/projected/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-kube-api-access-zw2n6\") pod \"infra-operator-controller-manager-5478cb7c68-h92m8\" (UID: \"8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9\") " pod="openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.233132 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjhlb\" (UniqueName: \"kubernetes.io/projected/8510a96b-51e8-4ad4-be1e-0fa7e2a86880-kube-api-access-bjhlb\") pod \"ironic-operator-controller-manager-554564d7fc-k28jp\" (UID: \"8510a96b-51e8-4ad4-be1e-0fa7e2a86880\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-k28jp" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.233152 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbjqp\" (UniqueName: \"kubernetes.io/projected/d4632ec1-624e-4097-9888-0bbe395d6916-kube-api-access-sbjqp\") pod \"horizon-operator-controller-manager-5b9b8895d5-m8p4r\" (UID: \"d4632ec1-624e-4097-9888-0bbe395d6916\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-m8p4r" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.236919 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-6hjt5"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.260173 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-hsx4c"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.260913 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-hsx4c" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.268900 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-hsx4c"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.282203 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-2bbz7" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.300162 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4xjq\" (UniqueName: \"kubernetes.io/projected/96c0bf26-aa73-4948-add8-b30fbbc23df9-kube-api-access-s4xjq\") pod \"heat-operator-controller-manager-69f49c598c-mbrvg\" (UID: \"96c0bf26-aa73-4948-add8-b30fbbc23df9\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-mbrvg" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.300820 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbjqp\" (UniqueName: \"kubernetes.io/projected/d4632ec1-624e-4097-9888-0bbe395d6916-kube-api-access-sbjqp\") pod \"horizon-operator-controller-manager-5b9b8895d5-m8p4r\" (UID: \"d4632ec1-624e-4097-9888-0bbe395d6916\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-m8p4r" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.322319 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-qd78d"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.323466 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-qd78d" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.332164 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-m8p4r" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.334008 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw2n6\" (UniqueName: \"kubernetes.io/projected/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-kube-api-access-zw2n6\") pod \"infra-operator-controller-manager-5478cb7c68-h92m8\" (UID: \"8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9\") " pod="openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.334067 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjhlb\" (UniqueName: \"kubernetes.io/projected/8510a96b-51e8-4ad4-be1e-0fa7e2a86880-kube-api-access-bjhlb\") pod \"ironic-operator-controller-manager-554564d7fc-k28jp\" (UID: \"8510a96b-51e8-4ad4-be1e-0fa7e2a86880\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-k28jp" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.334096 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wdn7\" (UniqueName: \"kubernetes.io/projected/56b8547b-7339-452b-9a06-a05411222f70-kube-api-access-7wdn7\") pod \"keystone-operator-controller-manager-b4d948c87-sjlh5\" (UID: \"56b8547b-7339-452b-9a06-a05411222f70\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-sjlh5" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.334132 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p5pm\" (UniqueName: \"kubernetes.io/projected/bac77172-b950-473c-8b07-0b1201a981db-kube-api-access-4p5pm\") pod \"neutron-operator-controller-manager-64ddbf8bb-hsx4c\" (UID: \"bac77172-b950-473c-8b07-0b1201a981db\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-hsx4c" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.334151 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4gfb\" (UniqueName: \"kubernetes.io/projected/23733893-aa4e-47e2-a06a-a055313308d8-kube-api-access-t4gfb\") pod \"manila-operator-controller-manager-54f6768c69-6hjt5\" (UID: \"23733893-aa4e-47e2-a06a-a055313308d8\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-6hjt5" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.334181 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-cert\") pod \"infra-operator-controller-manager-5478cb7c68-h92m8\" (UID: \"8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9\") " pod="openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8" Feb 14 11:23:26 crc kubenswrapper[4904]: E0214 11:23:26.334291 4904 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 14 11:23:26 crc kubenswrapper[4904]: E0214 11:23:26.334337 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-cert podName:8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:26.834319998 +0000 UTC m=+797.647084659 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-cert") pod "infra-operator-controller-manager-5478cb7c68-h92m8" (UID: "8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9") : secret "infra-operator-webhook-server-cert" not found Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.347305 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-bl8nv" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.373931 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-rvcg7"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.374983 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-rvcg7" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.380122 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-rvcg7"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.382827 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-zvs98" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.384511 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-qd78d"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.404749 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjhlb\" (UniqueName: \"kubernetes.io/projected/8510a96b-51e8-4ad4-be1e-0fa7e2a86880-kube-api-access-bjhlb\") pod \"ironic-operator-controller-manager-554564d7fc-k28jp\" (UID: \"8510a96b-51e8-4ad4-be1e-0fa7e2a86880\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-k28jp" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.404876 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw2n6\" (UniqueName: \"kubernetes.io/projected/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-kube-api-access-zw2n6\") pod \"infra-operator-controller-manager-5478cb7c68-h92m8\" (UID: \"8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9\") " pod="openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.410227 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-xb6dd"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.411577 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-xb6dd" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.415877 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.416816 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.422462 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.422766 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-kfc7g" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.422948 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-2t5j5" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.437629 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgfkx\" (UniqueName: \"kubernetes.io/projected/f29b166d-edb4-4403-be20-a523578848ee-kube-api-access-tgfkx\") pod \"nova-operator-controller-manager-567668f5cf-qd78d\" (UID: \"f29b166d-edb4-4403-be20-a523578848ee\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-qd78d" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.437733 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wdn7\" (UniqueName: \"kubernetes.io/projected/56b8547b-7339-452b-9a06-a05411222f70-kube-api-access-7wdn7\") pod \"keystone-operator-controller-manager-b4d948c87-sjlh5\" (UID: \"56b8547b-7339-452b-9a06-a05411222f70\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-sjlh5" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.437788 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p5pm\" (UniqueName: \"kubernetes.io/projected/bac77172-b950-473c-8b07-0b1201a981db-kube-api-access-4p5pm\") pod \"neutron-operator-controller-manager-64ddbf8bb-hsx4c\" (UID: \"bac77172-b950-473c-8b07-0b1201a981db\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-hsx4c" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.437808 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4gfb\" (UniqueName: \"kubernetes.io/projected/23733893-aa4e-47e2-a06a-a055313308d8-kube-api-access-t4gfb\") pod \"manila-operator-controller-manager-54f6768c69-6hjt5\" (UID: \"23733893-aa4e-47e2-a06a-a055313308d8\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-6hjt5" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.451392 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-9gwm7"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.452178 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-9gwm7" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.454580 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-f9kmn" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.472156 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wdn7\" (UniqueName: \"kubernetes.io/projected/56b8547b-7339-452b-9a06-a05411222f70-kube-api-access-7wdn7\") pod \"keystone-operator-controller-manager-b4d948c87-sjlh5\" (UID: \"56b8547b-7339-452b-9a06-a05411222f70\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-sjlh5" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.475622 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-pwsvp"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.476368 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-pwsvp" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.478795 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-tg2sc" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.503965 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-xb6dd"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.505100 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4gfb\" (UniqueName: \"kubernetes.io/projected/23733893-aa4e-47e2-a06a-a055313308d8-kube-api-access-t4gfb\") pod \"manila-operator-controller-manager-54f6768c69-6hjt5\" (UID: \"23733893-aa4e-47e2-a06a-a055313308d8\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-6hjt5" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.540791 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgfkx\" (UniqueName: \"kubernetes.io/projected/f29b166d-edb4-4403-be20-a523578848ee-kube-api-access-tgfkx\") pod \"nova-operator-controller-manager-567668f5cf-qd78d\" (UID: \"f29b166d-edb4-4403-be20-a523578848ee\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-qd78d" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.540874 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59gd8\" (UniqueName: \"kubernetes.io/projected/733ceff2-fece-49be-8bc4-63f3adcfff22-kube-api-access-59gd8\") pod \"ovn-operator-controller-manager-d44cf6b75-9gwm7\" (UID: \"733ceff2-fece-49be-8bc4-63f3adcfff22\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-9gwm7" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.540909 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5clf\" (UniqueName: \"kubernetes.io/projected/c3d1040e-0300-49ce-8b7d-7e4dac11c874-kube-api-access-z5clf\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c46s685\" (UID: \"c3d1040e-0300-49ce-8b7d-7e4dac11c874\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.540950 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tc7t\" (UniqueName: \"kubernetes.io/projected/cebffdc5-38fd-45a5-ab0a-7445696c1661-kube-api-access-5tc7t\") pod \"octavia-operator-controller-manager-69f8888797-xb6dd\" (UID: \"cebffdc5-38fd-45a5-ab0a-7445696c1661\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-xb6dd" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.540976 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f4mk\" (UniqueName: \"kubernetes.io/projected/70a8fe80-89cb-492b-a631-e8042e3ed1be-kube-api-access-4f4mk\") pod \"placement-operator-controller-manager-8497b45c89-pwsvp\" (UID: \"70a8fe80-89cb-492b-a631-e8042e3ed1be\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-pwsvp" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.575542 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p5pm\" (UniqueName: \"kubernetes.io/projected/bac77172-b950-473c-8b07-0b1201a981db-kube-api-access-4p5pm\") pod \"neutron-operator-controller-manager-64ddbf8bb-hsx4c\" (UID: \"bac77172-b950-473c-8b07-0b1201a981db\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-hsx4c" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.577244 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.585558 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3d1040e-0300-49ce-8b7d-7e4dac11c874-cert\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c46s685\" (UID: \"c3d1040e-0300-49ce-8b7d-7e4dac11c874\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.585624 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66fcr\" (UniqueName: \"kubernetes.io/projected/b80a5769-6d09-4697-ba97-a5723e161a3c-kube-api-access-66fcr\") pod \"mariadb-operator-controller-manager-6994f66f48-rvcg7\" (UID: \"b80a5769-6d09-4697-ba97-a5723e161a3c\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-rvcg7" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.623957 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-k28jp" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.624788 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-mbrvg" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.650760 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-4zj5b"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.654168 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-sjlh5" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.661292 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-4zj5b" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.664201 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgfkx\" (UniqueName: \"kubernetes.io/projected/f29b166d-edb4-4403-be20-a523578848ee-kube-api-access-tgfkx\") pod \"nova-operator-controller-manager-567668f5cf-qd78d\" (UID: \"f29b166d-edb4-4403-be20-a523578848ee\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-qd78d" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.666353 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-6fhs5" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.675326 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-6hjt5" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.693133 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3d1040e-0300-49ce-8b7d-7e4dac11c874-cert\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c46s685\" (UID: \"c3d1040e-0300-49ce-8b7d-7e4dac11c874\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.693186 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66fcr\" (UniqueName: \"kubernetes.io/projected/b80a5769-6d09-4697-ba97-a5723e161a3c-kube-api-access-66fcr\") pod \"mariadb-operator-controller-manager-6994f66f48-rvcg7\" (UID: \"b80a5769-6d09-4697-ba97-a5723e161a3c\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-rvcg7" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.693239 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59gd8\" (UniqueName: \"kubernetes.io/projected/733ceff2-fece-49be-8bc4-63f3adcfff22-kube-api-access-59gd8\") pod \"ovn-operator-controller-manager-d44cf6b75-9gwm7\" (UID: \"733ceff2-fece-49be-8bc4-63f3adcfff22\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-9gwm7" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.693264 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5clf\" (UniqueName: \"kubernetes.io/projected/c3d1040e-0300-49ce-8b7d-7e4dac11c874-kube-api-access-z5clf\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c46s685\" (UID: \"c3d1040e-0300-49ce-8b7d-7e4dac11c874\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.693323 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tc7t\" (UniqueName: \"kubernetes.io/projected/cebffdc5-38fd-45a5-ab0a-7445696c1661-kube-api-access-5tc7t\") pod \"octavia-operator-controller-manager-69f8888797-xb6dd\" (UID: \"cebffdc5-38fd-45a5-ab0a-7445696c1661\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-xb6dd" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.693349 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f4mk\" (UniqueName: \"kubernetes.io/projected/70a8fe80-89cb-492b-a631-e8042e3ed1be-kube-api-access-4f4mk\") pod \"placement-operator-controller-manager-8497b45c89-pwsvp\" (UID: \"70a8fe80-89cb-492b-a631-e8042e3ed1be\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-pwsvp" Feb 14 11:23:26 crc kubenswrapper[4904]: E0214 11:23:26.695532 4904 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 14 11:23:26 crc kubenswrapper[4904]: E0214 11:23:26.696192 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c3d1040e-0300-49ce-8b7d-7e4dac11c874-cert podName:c3d1040e-0300-49ce-8b7d-7e4dac11c874 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:27.196174487 +0000 UTC m=+798.008939148 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c3d1040e-0300-49ce-8b7d-7e4dac11c874-cert") pod "openstack-baremetal-operator-controller-manager-84966cf5c46s685" (UID: "c3d1040e-0300-49ce-8b7d-7e4dac11c874") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.704045 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-pwsvp"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.732089 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f4mk\" (UniqueName: \"kubernetes.io/projected/70a8fe80-89cb-492b-a631-e8042e3ed1be-kube-api-access-4f4mk\") pod \"placement-operator-controller-manager-8497b45c89-pwsvp\" (UID: \"70a8fe80-89cb-492b-a631-e8042e3ed1be\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-pwsvp" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.739456 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59gd8\" (UniqueName: \"kubernetes.io/projected/733ceff2-fece-49be-8bc4-63f3adcfff22-kube-api-access-59gd8\") pod \"ovn-operator-controller-manager-d44cf6b75-9gwm7\" (UID: \"733ceff2-fece-49be-8bc4-63f3adcfff22\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-9gwm7" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.739541 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-9gwm7"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.747074 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-hsx4c" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.765232 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q5cmd"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.766209 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q5cmd" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.767688 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-4zj5b"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.779886 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-qd78d" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.781039 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tc7t\" (UniqueName: \"kubernetes.io/projected/cebffdc5-38fd-45a5-ab0a-7445696c1661-kube-api-access-5tc7t\") pod \"octavia-operator-controller-manager-69f8888797-xb6dd\" (UID: \"cebffdc5-38fd-45a5-ab0a-7445696c1661\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-xb6dd" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.784815 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-z8rdm" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.792567 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5clf\" (UniqueName: \"kubernetes.io/projected/c3d1040e-0300-49ce-8b7d-7e4dac11c874-kube-api-access-z5clf\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c46s685\" (UID: \"c3d1040e-0300-49ce-8b7d-7e4dac11c874\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.794508 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82gcc\" (UniqueName: \"kubernetes.io/projected/fa0069c3-fd8a-4ffe-aad0-ccee78f70caa-kube-api-access-82gcc\") pod \"swift-operator-controller-manager-68f46476f-4zj5b\" (UID: \"fa0069c3-fd8a-4ffe-aad0-ccee78f70caa\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-4zj5b" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.794992 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66fcr\" (UniqueName: \"kubernetes.io/projected/b80a5769-6d09-4697-ba97-a5723e161a3c-kube-api-access-66fcr\") pod \"mariadb-operator-controller-manager-6994f66f48-rvcg7\" (UID: \"b80a5769-6d09-4697-ba97-a5723e161a3c\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-rvcg7" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.798747 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q5cmd"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.812526 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-rvcg7" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.819066 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-jnd94"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.819989 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7866795846-jnd94" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.826828 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-2qq5k" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.830995 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-xb6dd" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.880003 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-x8k6m"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.880901 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x8k6m" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.884555 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-8bbrr" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.890801 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-9gwm7" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.895708 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpzg8\" (UniqueName: \"kubernetes.io/projected/95a68c48-4e90-43c0-9caf-d6be291fca1e-kube-api-access-jpzg8\") pod \"telemetry-operator-controller-manager-7f45b4ff68-q5cmd\" (UID: \"95a68c48-4e90-43c0-9caf-d6be291fca1e\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q5cmd" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.895754 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-cert\") pod \"infra-operator-controller-manager-5478cb7c68-h92m8\" (UID: \"8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9\") " pod="openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.895810 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82gcc\" (UniqueName: \"kubernetes.io/projected/fa0069c3-fd8a-4ffe-aad0-ccee78f70caa-kube-api-access-82gcc\") pod \"swift-operator-controller-manager-68f46476f-4zj5b\" (UID: \"fa0069c3-fd8a-4ffe-aad0-ccee78f70caa\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-4zj5b" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.895912 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ft2p\" (UniqueName: \"kubernetes.io/projected/00d1bd7d-c3d8-4a54-9208-c28355796751-kube-api-access-4ft2p\") pod \"test-operator-controller-manager-7866795846-jnd94\" (UID: \"00d1bd7d-c3d8-4a54-9208-c28355796751\") " pod="openstack-operators/test-operator-controller-manager-7866795846-jnd94" Feb 14 11:23:26 crc kubenswrapper[4904]: E0214 11:23:26.896706 4904 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 14 11:23:26 crc kubenswrapper[4904]: E0214 11:23:26.896749 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-cert podName:8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:27.896736863 +0000 UTC m=+798.709501524 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-cert") pod "infra-operator-controller-manager-5478cb7c68-h92m8" (UID: "8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9") : secret "infra-operator-webhook-server-cert" not found Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.900105 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-jnd94"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.929138 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-pwsvp" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.961570 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-x8k6m"] Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.963725 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82gcc\" (UniqueName: \"kubernetes.io/projected/fa0069c3-fd8a-4ffe-aad0-ccee78f70caa-kube-api-access-82gcc\") pod \"swift-operator-controller-manager-68f46476f-4zj5b\" (UID: \"fa0069c3-fd8a-4ffe-aad0-ccee78f70caa\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-4zj5b" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.998415 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ft2p\" (UniqueName: \"kubernetes.io/projected/00d1bd7d-c3d8-4a54-9208-c28355796751-kube-api-access-4ft2p\") pod \"test-operator-controller-manager-7866795846-jnd94\" (UID: \"00d1bd7d-c3d8-4a54-9208-c28355796751\") " pod="openstack-operators/test-operator-controller-manager-7866795846-jnd94" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.998516 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpzg8\" (UniqueName: \"kubernetes.io/projected/95a68c48-4e90-43c0-9caf-d6be291fca1e-kube-api-access-jpzg8\") pod \"telemetry-operator-controller-manager-7f45b4ff68-q5cmd\" (UID: \"95a68c48-4e90-43c0-9caf-d6be291fca1e\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q5cmd" Feb 14 11:23:26 crc kubenswrapper[4904]: I0214 11:23:26.998552 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knpzr\" (UniqueName: \"kubernetes.io/projected/53ba6432-a3b1-4cce-8ab2-bc74a718b134-kube-api-access-knpzr\") pod \"watcher-operator-controller-manager-5db88f68c-x8k6m\" (UID: \"53ba6432-a3b1-4cce-8ab2-bc74a718b134\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x8k6m" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.017696 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-4zj5b" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.027450 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpzg8\" (UniqueName: \"kubernetes.io/projected/95a68c48-4e90-43c0-9caf-d6be291fca1e-kube-api-access-jpzg8\") pod \"telemetry-operator-controller-manager-7f45b4ff68-q5cmd\" (UID: \"95a68c48-4e90-43c0-9caf-d6be291fca1e\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q5cmd" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.051183 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ft2p\" (UniqueName: \"kubernetes.io/projected/00d1bd7d-c3d8-4a54-9208-c28355796751-kube-api-access-4ft2p\") pod \"test-operator-controller-manager-7866795846-jnd94\" (UID: \"00d1bd7d-c3d8-4a54-9208-c28355796751\") " pod="openstack-operators/test-operator-controller-manager-7866795846-jnd94" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.055944 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7"] Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.060646 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.068506 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.069089 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.069453 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-s9nm5" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.074673 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7"] Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.102027 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knpzr\" (UniqueName: \"kubernetes.io/projected/53ba6432-a3b1-4cce-8ab2-bc74a718b134-kube-api-access-knpzr\") pod \"watcher-operator-controller-manager-5db88f68c-x8k6m\" (UID: \"53ba6432-a3b1-4cce-8ab2-bc74a718b134\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x8k6m" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.102108 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cvhx\" (UniqueName: \"kubernetes.io/projected/850dfcee-db10-493e-8043-5e167cfe3a54-kube-api-access-8cvhx\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.102162 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.102231 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.102741 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q5cmd" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.129883 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-795l4"] Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.130699 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-795l4" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.133199 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-h2s4m" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.134858 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knpzr\" (UniqueName: \"kubernetes.io/projected/53ba6432-a3b1-4cce-8ab2-bc74a718b134-kube-api-access-knpzr\") pod \"watcher-operator-controller-manager-5db88f68c-x8k6m\" (UID: \"53ba6432-a3b1-4cce-8ab2-bc74a718b134\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x8k6m" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.134917 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-795l4"] Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.176270 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7866795846-jnd94" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.204898 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cvhx\" (UniqueName: \"kubernetes.io/projected/850dfcee-db10-493e-8043-5e167cfe3a54-kube-api-access-8cvhx\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.204956 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.205006 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3d1040e-0300-49ce-8b7d-7e4dac11c874-cert\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c46s685\" (UID: \"c3d1040e-0300-49ce-8b7d-7e4dac11c874\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.205023 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.205050 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skl4f\" (UniqueName: \"kubernetes.io/projected/aeb77a69-e9d9-4e04-a270-665a28bb0fa1-kube-api-access-skl4f\") pod \"rabbitmq-cluster-operator-manager-668c99d594-795l4\" (UID: \"aeb77a69-e9d9-4e04-a270-665a28bb0fa1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-795l4" Feb 14 11:23:27 crc kubenswrapper[4904]: E0214 11:23:27.205407 4904 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 14 11:23:27 crc kubenswrapper[4904]: E0214 11:23:27.205442 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs podName:850dfcee-db10-493e-8043-5e167cfe3a54 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:27.705430873 +0000 UTC m=+798.518195534 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs") pod "openstack-operator-controller-manager-67d96975f-hzkn7" (UID: "850dfcee-db10-493e-8043-5e167cfe3a54") : secret "metrics-server-cert" not found Feb 14 11:23:27 crc kubenswrapper[4904]: E0214 11:23:27.205581 4904 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 14 11:23:27 crc kubenswrapper[4904]: E0214 11:23:27.205610 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c3d1040e-0300-49ce-8b7d-7e4dac11c874-cert podName:c3d1040e-0300-49ce-8b7d-7e4dac11c874 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:28.205601648 +0000 UTC m=+799.018366309 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c3d1040e-0300-49ce-8b7d-7e4dac11c874-cert") pod "openstack-baremetal-operator-controller-manager-84966cf5c46s685" (UID: "c3d1040e-0300-49ce-8b7d-7e4dac11c874") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 14 11:23:27 crc kubenswrapper[4904]: E0214 11:23:27.205649 4904 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 14 11:23:27 crc kubenswrapper[4904]: E0214 11:23:27.205667 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs podName:850dfcee-db10-493e-8043-5e167cfe3a54 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:27.70566106 +0000 UTC m=+798.518425721 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs") pod "openstack-operator-controller-manager-67d96975f-hzkn7" (UID: "850dfcee-db10-493e-8043-5e167cfe3a54") : secret "webhook-server-cert" not found Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.218701 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-hbqp9"] Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.219217 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x8k6m" Feb 14 11:23:27 crc kubenswrapper[4904]: W0214 11:23:27.236966 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b26d95c_682d_4be5_9788_5044e186611c.slice/crio-2313fe579d46906d4459fa9873c514bf7b61f74ec3653cbd4aabc89a139c43b9 WatchSource:0}: Error finding container 2313fe579d46906d4459fa9873c514bf7b61f74ec3653cbd4aabc89a139c43b9: Status 404 returned error can't find the container with id 2313fe579d46906d4459fa9873c514bf7b61f74ec3653cbd4aabc89a139c43b9 Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.237950 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-m29fl"] Feb 14 11:23:27 crc kubenswrapper[4904]: W0214 11:23:27.253553 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33670c4d_fd2a_49cd_b602_ccdd927b7c00.slice/crio-78d232cbde3c3e5113f2a382df07086fcaf20a0266d9d66ed57bfa09ed116ab5 WatchSource:0}: Error finding container 78d232cbde3c3e5113f2a382df07086fcaf20a0266d9d66ed57bfa09ed116ab5: Status 404 returned error can't find the container with id 78d232cbde3c3e5113f2a382df07086fcaf20a0266d9d66ed57bfa09ed116ab5 Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.266711 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cvhx\" (UniqueName: \"kubernetes.io/projected/850dfcee-db10-493e-8043-5e167cfe3a54-kube-api-access-8cvhx\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.308579 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skl4f\" (UniqueName: \"kubernetes.io/projected/aeb77a69-e9d9-4e04-a270-665a28bb0fa1-kube-api-access-skl4f\") pod \"rabbitmq-cluster-operator-manager-668c99d594-795l4\" (UID: \"aeb77a69-e9d9-4e04-a270-665a28bb0fa1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-795l4" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.351365 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skl4f\" (UniqueName: \"kubernetes.io/projected/aeb77a69-e9d9-4e04-a270-665a28bb0fa1-kube-api-access-skl4f\") pod \"rabbitmq-cluster-operator-manager-668c99d594-795l4\" (UID: \"aeb77a69-e9d9-4e04-a270-665a28bb0fa1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-795l4" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.452637 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-gt547"] Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.504023 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-jsn2d"] Feb 14 11:23:27 crc kubenswrapper[4904]: W0214 11:23:27.547171 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2db31100_df1f_441c_87bf_944dc247a4fc.slice/crio-9cb9965687b69a64ed4c7c93554052525cb6960b42b34b4b3cfacb91f7c73387 WatchSource:0}: Error finding container 9cb9965687b69a64ed4c7c93554052525cb6960b42b34b4b3cfacb91f7c73387: Status 404 returned error can't find the container with id 9cb9965687b69a64ed4c7c93554052525cb6960b42b34b4b3cfacb91f7c73387 Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.587409 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-795l4" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.590962 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-m8p4r"] Feb 14 11:23:27 crc kubenswrapper[4904]: W0214 11:23:27.612547 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4632ec1_624e_4097_9888_0bbe395d6916.slice/crio-9909c455a00f4e4d54d9682ec97cfdb489266474933d4d0fe6d57627b1e5c8cf WatchSource:0}: Error finding container 9909c455a00f4e4d54d9682ec97cfdb489266474933d4d0fe6d57627b1e5c8cf: Status 404 returned error can't find the container with id 9909c455a00f4e4d54d9682ec97cfdb489266474933d4d0fe6d57627b1e5c8cf Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.674293 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-mbrvg"] Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.719595 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.719729 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:27 crc kubenswrapper[4904]: E0214 11:23:27.719920 4904 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 14 11:23:27 crc kubenswrapper[4904]: E0214 11:23:27.720010 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs podName:850dfcee-db10-493e-8043-5e167cfe3a54 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:28.719987757 +0000 UTC m=+799.532752488 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs") pod "openstack-operator-controller-manager-67d96975f-hzkn7" (UID: "850dfcee-db10-493e-8043-5e167cfe3a54") : secret "webhook-server-cert" not found Feb 14 11:23:27 crc kubenswrapper[4904]: E0214 11:23:27.720388 4904 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 14 11:23:27 crc kubenswrapper[4904]: E0214 11:23:27.720450 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs podName:850dfcee-db10-493e-8043-5e167cfe3a54 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:28.720437939 +0000 UTC m=+799.533202680 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs") pod "openstack-operator-controller-manager-67d96975f-hzkn7" (UID: "850dfcee-db10-493e-8043-5e167cfe3a54") : secret "metrics-server-cert" not found Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.908218 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-sjlh5"] Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.919599 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-qd78d"] Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.926828 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-cert\") pod \"infra-operator-controller-manager-5478cb7c68-h92m8\" (UID: \"8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9\") " pod="openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8" Feb 14 11:23:27 crc kubenswrapper[4904]: E0214 11:23:27.927094 4904 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 14 11:23:27 crc kubenswrapper[4904]: E0214 11:23:27.927612 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-cert podName:8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:29.927591048 +0000 UTC m=+800.740355709 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-cert") pod "infra-operator-controller-manager-5478cb7c68-h92m8" (UID: "8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9") : secret "infra-operator-webhook-server-cert" not found Feb 14 11:23:27 crc kubenswrapper[4904]: I0214 11:23:27.931973 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-k28jp"] Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.169566 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-rvcg7"] Feb 14 11:23:28 crc kubenswrapper[4904]: W0214 11:23:28.185646 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbac77172_b950_473c_8b07_0b1201a981db.slice/crio-0223ad9a6e536fba7fbc0b3f4cdc9f21bc9a3b42127e64c7b3b52f33f95f3cbf WatchSource:0}: Error finding container 0223ad9a6e536fba7fbc0b3f4cdc9f21bc9a3b42127e64c7b3b52f33f95f3cbf: Status 404 returned error can't find the container with id 0223ad9a6e536fba7fbc0b3f4cdc9f21bc9a3b42127e64c7b3b52f33f95f3cbf Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.199247 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-6hjt5"] Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.232760 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3d1040e-0300-49ce-8b7d-7e4dac11c874-cert\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c46s685\" (UID: \"c3d1040e-0300-49ce-8b7d-7e4dac11c874\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685" Feb 14 11:23:28 crc kubenswrapper[4904]: E0214 11:23:28.232971 4904 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 14 11:23:28 crc kubenswrapper[4904]: E0214 11:23:28.233045 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c3d1040e-0300-49ce-8b7d-7e4dac11c874-cert podName:c3d1040e-0300-49ce-8b7d-7e4dac11c874 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:30.233024188 +0000 UTC m=+801.045788849 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c3d1040e-0300-49ce-8b7d-7e4dac11c874-cert") pod "openstack-baremetal-operator-controller-manager-84966cf5c46s685" (UID: "c3d1040e-0300-49ce-8b7d-7e4dac11c874") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.246562 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-hsx4c"] Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.309856 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-gt547" event={"ID":"6c57743e-c864-4e0d-828a-406978aab771","Type":"ContainerStarted","Data":"762248902bd20ddf326362bab78b9d10df96fa9110c54a78d39da8e36d68545b"} Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.312192 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-xb6dd"] Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.314337 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-6hjt5" event={"ID":"23733893-aa4e-47e2-a06a-a055313308d8","Type":"ContainerStarted","Data":"b0c962a3eb2161534bbd35621edce412cc179899c6d1f72f2145fc9bf6866712"} Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.322603 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-9gwm7"] Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.323946 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-qd78d" event={"ID":"f29b166d-edb4-4403-be20-a523578848ee","Type":"ContainerStarted","Data":"32e78d67706e717d7303d07a3cadf11bcaafc687c020844374a45ba976d1cb35"} Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.325188 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-k28jp" event={"ID":"8510a96b-51e8-4ad4-be1e-0fa7e2a86880","Type":"ContainerStarted","Data":"e08ea21a5e8213fcb2d9fc30f8b2611543952100cbf2d21b7c92a40a92e21254"} Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.326549 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987464f4-jsn2d" event={"ID":"2db31100-df1f-441c-87bf-944dc247a4fc","Type":"ContainerStarted","Data":"9cb9965687b69a64ed4c7c93554052525cb6960b42b34b4b3cfacb91f7c73387"} Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.327415 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-mbrvg" event={"ID":"96c0bf26-aa73-4948-add8-b30fbbc23df9","Type":"ContainerStarted","Data":"7c5a544b44c9050666a00f727f51b5367f7c07ff0583aa7d6cc63676cbb61d6b"} Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.340555 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-hsx4c" event={"ID":"bac77172-b950-473c-8b07-0b1201a981db","Type":"ContainerStarted","Data":"0223ad9a6e536fba7fbc0b3f4cdc9f21bc9a3b42127e64c7b3b52f33f95f3cbf"} Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.351367 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-m8p4r" event={"ID":"d4632ec1-624e-4097-9888-0bbe395d6916","Type":"ContainerStarted","Data":"9909c455a00f4e4d54d9682ec97cfdb489266474933d4d0fe6d57627b1e5c8cf"} Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.352826 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-hbqp9" event={"ID":"33670c4d-fd2a-49cd-b602-ccdd927b7c00","Type":"ContainerStarted","Data":"78d232cbde3c3e5113f2a382df07086fcaf20a0266d9d66ed57bfa09ed116ab5"} Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.354256 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-sjlh5" event={"ID":"56b8547b-7339-452b-9a06-a05411222f70","Type":"ContainerStarted","Data":"60c4fd0b61cd40bbac5c8eb0924f65a5008c9f544c82107187bbdc6f956db4a1"} Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.357920 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-pwsvp"] Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.377015 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-rvcg7" event={"ID":"b80a5769-6d09-4697-ba97-a5723e161a3c","Type":"ContainerStarted","Data":"a5b09e108e256b8a62aefbc0cfc0928155060f6d1f464f5ebc963a02a3cf5cb9"} Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.393887 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-m29fl" event={"ID":"2b26d95c-682d-4be5-9788-5044e186611c","Type":"ContainerStarted","Data":"2313fe579d46906d4459fa9873c514bf7b61f74ec3653cbd4aabc89a139c43b9"} Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.492221 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-4zj5b"] Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.524921 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-jnd94"] Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.547036 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-795l4"] Feb 14 11:23:28 crc kubenswrapper[4904]: E0214 11:23:28.581806 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-skl4f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-795l4_openstack-operators(aeb77a69-e9d9-4e04-a270-665a28bb0fa1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 14 11:23:28 crc kubenswrapper[4904]: E0214 11:23:28.583112 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-795l4" podUID="aeb77a69-e9d9-4e04-a270-665a28bb0fa1" Feb 14 11:23:28 crc kubenswrapper[4904]: W0214 11:23:28.587511 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00d1bd7d_c3d8_4a54_9208_c28355796751.slice/crio-ef8c8e453fd23b6496e77c3e5e5155b4e66a3e61adb30cae1a7f43adb70eba5b WatchSource:0}: Error finding container ef8c8e453fd23b6496e77c3e5e5155b4e66a3e61adb30cae1a7f43adb70eba5b: Status 404 returned error can't find the container with id ef8c8e453fd23b6496e77c3e5e5155b4e66a3e61adb30cae1a7f43adb70eba5b Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.601069 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q5cmd"] Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.601120 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-x8k6m"] Feb 14 11:23:28 crc kubenswrapper[4904]: E0214 11:23:28.614528 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4ft2p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-7866795846-jnd94_openstack-operators(00d1bd7d-c3d8-4a54-9208-c28355796751): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 14 11:23:28 crc kubenswrapper[4904]: E0214 11:23:28.615820 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-7866795846-jnd94" podUID="00d1bd7d-c3d8-4a54-9208-c28355796751" Feb 14 11:23:28 crc kubenswrapper[4904]: E0214 11:23:28.632621 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:66a4b9322ebb573313178ea88e31026d4532f461592b9fae2dff71efd9256d99,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jpzg8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-7f45b4ff68-q5cmd_openstack-operators(95a68c48-4e90-43c0-9caf-d6be291fca1e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 14 11:23:28 crc kubenswrapper[4904]: E0214 11:23:28.633530 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:d01ae848290e880c09127d5297418dea40fc7f090fdab9bf2c578c7e7f53aec0,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-knpzr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-5db88f68c-x8k6m_openstack-operators(53ba6432-a3b1-4cce-8ab2-bc74a718b134): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 14 11:23:28 crc kubenswrapper[4904]: E0214 11:23:28.636516 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x8k6m" podUID="53ba6432-a3b1-4cce-8ab2-bc74a718b134" Feb 14 11:23:28 crc kubenswrapper[4904]: E0214 11:23:28.636770 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q5cmd" podUID="95a68c48-4e90-43c0-9caf-d6be291fca1e" Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.752739 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:28 crc kubenswrapper[4904]: I0214 11:23:28.752889 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:28 crc kubenswrapper[4904]: E0214 11:23:28.753145 4904 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 14 11:23:28 crc kubenswrapper[4904]: E0214 11:23:28.753209 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs podName:850dfcee-db10-493e-8043-5e167cfe3a54 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:30.753189627 +0000 UTC m=+801.565954288 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs") pod "openstack-operator-controller-manager-67d96975f-hzkn7" (UID: "850dfcee-db10-493e-8043-5e167cfe3a54") : secret "webhook-server-cert" not found Feb 14 11:23:28 crc kubenswrapper[4904]: E0214 11:23:28.753657 4904 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 14 11:23:28 crc kubenswrapper[4904]: E0214 11:23:28.753703 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs podName:850dfcee-db10-493e-8043-5e167cfe3a54 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:30.75369071 +0000 UTC m=+801.566455381 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs") pod "openstack-operator-controller-manager-67d96975f-hzkn7" (UID: "850dfcee-db10-493e-8043-5e167cfe3a54") : secret "metrics-server-cert" not found Feb 14 11:23:29 crc kubenswrapper[4904]: I0214 11:23:29.411729 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-4zj5b" event={"ID":"fa0069c3-fd8a-4ffe-aad0-ccee78f70caa","Type":"ContainerStarted","Data":"2c90ee09552e420204f9d1548e16187e2ab345c822210c080a83ca81488a990c"} Feb 14 11:23:29 crc kubenswrapper[4904]: I0214 11:23:29.425775 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-9gwm7" event={"ID":"733ceff2-fece-49be-8bc4-63f3adcfff22","Type":"ContainerStarted","Data":"c66291f082c812130b7288d1c342fcfadf07a4976ede1f1ceac85259eb6decd6"} Feb 14 11:23:29 crc kubenswrapper[4904]: I0214 11:23:29.427967 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-xb6dd" event={"ID":"cebffdc5-38fd-45a5-ab0a-7445696c1661","Type":"ContainerStarted","Data":"8117660e3aced287914c6697c8c71010fdc04a750a73db3895cc4771b4578156"} Feb 14 11:23:29 crc kubenswrapper[4904]: I0214 11:23:29.430431 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x8k6m" event={"ID":"53ba6432-a3b1-4cce-8ab2-bc74a718b134","Type":"ContainerStarted","Data":"7b16a9cbb92c6dabcaf08c5dc7bbf2140f7ee83e070123f629ae6eb3ef4b86c4"} Feb 14 11:23:29 crc kubenswrapper[4904]: E0214 11:23:29.439603 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d01ae848290e880c09127d5297418dea40fc7f090fdab9bf2c578c7e7f53aec0\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x8k6m" podUID="53ba6432-a3b1-4cce-8ab2-bc74a718b134" Feb 14 11:23:29 crc kubenswrapper[4904]: I0214 11:23:29.442703 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-pwsvp" event={"ID":"70a8fe80-89cb-492b-a631-e8042e3ed1be","Type":"ContainerStarted","Data":"e6b4302527f061a2b7c0aa5242476a924d21944e22f6eae9560cf7a20e34c219"} Feb 14 11:23:29 crc kubenswrapper[4904]: I0214 11:23:29.448961 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q5cmd" event={"ID":"95a68c48-4e90-43c0-9caf-d6be291fca1e","Type":"ContainerStarted","Data":"5d0154560a0b83254430a8dcba7f8d3dde4159992690a9ffe5f08610c717fdc3"} Feb 14 11:23:29 crc kubenswrapper[4904]: E0214 11:23:29.454859 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:66a4b9322ebb573313178ea88e31026d4532f461592b9fae2dff71efd9256d99\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q5cmd" podUID="95a68c48-4e90-43c0-9caf-d6be291fca1e" Feb 14 11:23:29 crc kubenswrapper[4904]: I0214 11:23:29.455619 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-795l4" event={"ID":"aeb77a69-e9d9-4e04-a270-665a28bb0fa1","Type":"ContainerStarted","Data":"5892aed57f2e15ce2f9a6bd0207bbf98cdd043d15af8006a38587873ef3a7f0c"} Feb 14 11:23:29 crc kubenswrapper[4904]: E0214 11:23:29.456821 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-795l4" podUID="aeb77a69-e9d9-4e04-a270-665a28bb0fa1" Feb 14 11:23:29 crc kubenswrapper[4904]: I0214 11:23:29.459076 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7866795846-jnd94" event={"ID":"00d1bd7d-c3d8-4a54-9208-c28355796751","Type":"ContainerStarted","Data":"ef8c8e453fd23b6496e77c3e5e5155b4e66a3e61adb30cae1a7f43adb70eba5b"} Feb 14 11:23:29 crc kubenswrapper[4904]: E0214 11:23:29.463642 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6\\\"\"" pod="openstack-operators/test-operator-controller-manager-7866795846-jnd94" podUID="00d1bd7d-c3d8-4a54-9208-c28355796751" Feb 14 11:23:29 crc kubenswrapper[4904]: I0214 11:23:29.980031 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-cert\") pod \"infra-operator-controller-manager-5478cb7c68-h92m8\" (UID: \"8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9\") " pod="openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8" Feb 14 11:23:29 crc kubenswrapper[4904]: E0214 11:23:29.981269 4904 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 14 11:23:29 crc kubenswrapper[4904]: E0214 11:23:29.981333 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-cert podName:8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:33.981316217 +0000 UTC m=+804.794080888 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-cert") pod "infra-operator-controller-manager-5478cb7c68-h92m8" (UID: "8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9") : secret "infra-operator-webhook-server-cert" not found Feb 14 11:23:30 crc kubenswrapper[4904]: I0214 11:23:30.284949 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3d1040e-0300-49ce-8b7d-7e4dac11c874-cert\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c46s685\" (UID: \"c3d1040e-0300-49ce-8b7d-7e4dac11c874\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685" Feb 14 11:23:30 crc kubenswrapper[4904]: E0214 11:23:30.285076 4904 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 14 11:23:30 crc kubenswrapper[4904]: E0214 11:23:30.285134 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c3d1040e-0300-49ce-8b7d-7e4dac11c874-cert podName:c3d1040e-0300-49ce-8b7d-7e4dac11c874 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:34.285119113 +0000 UTC m=+805.097883774 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c3d1040e-0300-49ce-8b7d-7e4dac11c874-cert") pod "openstack-baremetal-operator-controller-manager-84966cf5c46s685" (UID: "c3d1040e-0300-49ce-8b7d-7e4dac11c874") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 14 11:23:30 crc kubenswrapper[4904]: E0214 11:23:30.481025 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:66a4b9322ebb573313178ea88e31026d4532f461592b9fae2dff71efd9256d99\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q5cmd" podUID="95a68c48-4e90-43c0-9caf-d6be291fca1e" Feb 14 11:23:30 crc kubenswrapper[4904]: E0214 11:23:30.483112 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-795l4" podUID="aeb77a69-e9d9-4e04-a270-665a28bb0fa1" Feb 14 11:23:30 crc kubenswrapper[4904]: E0214 11:23:30.483176 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d01ae848290e880c09127d5297418dea40fc7f090fdab9bf2c578c7e7f53aec0\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x8k6m" podUID="53ba6432-a3b1-4cce-8ab2-bc74a718b134" Feb 14 11:23:30 crc kubenswrapper[4904]: E0214 11:23:30.483232 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6\\\"\"" pod="openstack-operators/test-operator-controller-manager-7866795846-jnd94" podUID="00d1bd7d-c3d8-4a54-9208-c28355796751" Feb 14 11:23:30 crc kubenswrapper[4904]: I0214 11:23:30.796356 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:30 crc kubenswrapper[4904]: I0214 11:23:30.796427 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:30 crc kubenswrapper[4904]: E0214 11:23:30.796558 4904 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 14 11:23:30 crc kubenswrapper[4904]: E0214 11:23:30.796610 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs podName:850dfcee-db10-493e-8043-5e167cfe3a54 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:34.796595811 +0000 UTC m=+805.609360472 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs") pod "openstack-operator-controller-manager-67d96975f-hzkn7" (UID: "850dfcee-db10-493e-8043-5e167cfe3a54") : secret "webhook-server-cert" not found Feb 14 11:23:30 crc kubenswrapper[4904]: E0214 11:23:30.796912 4904 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 14 11:23:30 crc kubenswrapper[4904]: E0214 11:23:30.796942 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs podName:850dfcee-db10-493e-8043-5e167cfe3a54 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:34.79693538 +0000 UTC m=+805.609700041 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs") pod "openstack-operator-controller-manager-67d96975f-hzkn7" (UID: "850dfcee-db10-493e-8043-5e167cfe3a54") : secret "metrics-server-cert" not found Feb 14 11:23:34 crc kubenswrapper[4904]: I0214 11:23:34.052293 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-cert\") pod \"infra-operator-controller-manager-5478cb7c68-h92m8\" (UID: \"8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9\") " pod="openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8" Feb 14 11:23:34 crc kubenswrapper[4904]: E0214 11:23:34.052658 4904 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 14 11:23:34 crc kubenswrapper[4904]: E0214 11:23:34.052738 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-cert podName:8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:42.05272039 +0000 UTC m=+812.865485051 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-cert") pod "infra-operator-controller-manager-5478cb7c68-h92m8" (UID: "8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9") : secret "infra-operator-webhook-server-cert" not found Feb 14 11:23:34 crc kubenswrapper[4904]: I0214 11:23:34.356166 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3d1040e-0300-49ce-8b7d-7e4dac11c874-cert\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c46s685\" (UID: \"c3d1040e-0300-49ce-8b7d-7e4dac11c874\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685" Feb 14 11:23:34 crc kubenswrapper[4904]: E0214 11:23:34.356367 4904 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 14 11:23:34 crc kubenswrapper[4904]: E0214 11:23:34.356520 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c3d1040e-0300-49ce-8b7d-7e4dac11c874-cert podName:c3d1040e-0300-49ce-8b7d-7e4dac11c874 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:42.356504706 +0000 UTC m=+813.169269367 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c3d1040e-0300-49ce-8b7d-7e4dac11c874-cert") pod "openstack-baremetal-operator-controller-manager-84966cf5c46s685" (UID: "c3d1040e-0300-49ce-8b7d-7e4dac11c874") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 14 11:23:34 crc kubenswrapper[4904]: I0214 11:23:34.864052 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:34 crc kubenswrapper[4904]: I0214 11:23:34.864152 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:34 crc kubenswrapper[4904]: E0214 11:23:34.864277 4904 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 14 11:23:34 crc kubenswrapper[4904]: E0214 11:23:34.864327 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs podName:850dfcee-db10-493e-8043-5e167cfe3a54 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:42.864313772 +0000 UTC m=+813.677078433 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs") pod "openstack-operator-controller-manager-67d96975f-hzkn7" (UID: "850dfcee-db10-493e-8043-5e167cfe3a54") : secret "webhook-server-cert" not found Feb 14 11:23:34 crc kubenswrapper[4904]: E0214 11:23:34.864625 4904 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 14 11:23:34 crc kubenswrapper[4904]: E0214 11:23:34.864653 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs podName:850dfcee-db10-493e-8043-5e167cfe3a54 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:42.864646552 +0000 UTC m=+813.677411213 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs") pod "openstack-operator-controller-manager-67d96975f-hzkn7" (UID: "850dfcee-db10-493e-8043-5e167cfe3a54") : secret "metrics-server-cert" not found Feb 14 11:23:42 crc kubenswrapper[4904]: I0214 11:23:42.073640 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-cert\") pod \"infra-operator-controller-manager-5478cb7c68-h92m8\" (UID: \"8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9\") " pod="openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8" Feb 14 11:23:42 crc kubenswrapper[4904]: I0214 11:23:42.079350 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9-cert\") pod \"infra-operator-controller-manager-5478cb7c68-h92m8\" (UID: \"8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9\") " pod="openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8" Feb 14 11:23:42 crc kubenswrapper[4904]: I0214 11:23:42.132917 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8" Feb 14 11:23:42 crc kubenswrapper[4904]: I0214 11:23:42.378663 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3d1040e-0300-49ce-8b7d-7e4dac11c874-cert\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c46s685\" (UID: \"c3d1040e-0300-49ce-8b7d-7e4dac11c874\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685" Feb 14 11:23:42 crc kubenswrapper[4904]: I0214 11:23:42.383457 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3d1040e-0300-49ce-8b7d-7e4dac11c874-cert\") pod \"openstack-baremetal-operator-controller-manager-84966cf5c46s685\" (UID: \"c3d1040e-0300-49ce-8b7d-7e4dac11c874\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685" Feb 14 11:23:42 crc kubenswrapper[4904]: I0214 11:23:42.477268 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685" Feb 14 11:23:42 crc kubenswrapper[4904]: I0214 11:23:42.885028 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:42 crc kubenswrapper[4904]: E0214 11:23:42.885176 4904 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 14 11:23:42 crc kubenswrapper[4904]: E0214 11:23:42.885443 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs podName:850dfcee-db10-493e-8043-5e167cfe3a54 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:58.88542609 +0000 UTC m=+829.698190751 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs") pod "openstack-operator-controller-manager-67d96975f-hzkn7" (UID: "850dfcee-db10-493e-8043-5e167cfe3a54") : secret "webhook-server-cert" not found Feb 14 11:23:42 crc kubenswrapper[4904]: I0214 11:23:42.885629 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:42 crc kubenswrapper[4904]: E0214 11:23:42.886075 4904 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 14 11:23:42 crc kubenswrapper[4904]: E0214 11:23:42.886106 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs podName:850dfcee-db10-493e-8043-5e167cfe3a54 nodeName:}" failed. No retries permitted until 2026-02-14 11:23:58.886098049 +0000 UTC m=+829.698862710 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs") pod "openstack-operator-controller-manager-67d96975f-hzkn7" (UID: "850dfcee-db10-493e-8043-5e167cfe3a54") : secret "metrics-server-cert" not found Feb 14 11:23:42 crc kubenswrapper[4904]: E0214 11:23:42.937064 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:543c103838f3e6ef48755665a7695dfa3ed84753c557560257d265db31f92759" Feb 14 11:23:42 crc kubenswrapper[4904]: E0214 11:23:42.937258 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:543c103838f3e6ef48755665a7695dfa3ed84753c557560257d265db31f92759,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-59gd8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-d44cf6b75-9gwm7_openstack-operators(733ceff2-fece-49be-8bc4-63f3adcfff22): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:23:42 crc kubenswrapper[4904]: E0214 11:23:42.938452 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-9gwm7" podUID="733ceff2-fece-49be-8bc4-63f3adcfff22" Feb 14 11:23:43 crc kubenswrapper[4904]: E0214 11:23:43.571308 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:543c103838f3e6ef48755665a7695dfa3ed84753c557560257d265db31f92759\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-9gwm7" podUID="733ceff2-fece-49be-8bc4-63f3adcfff22" Feb 14 11:23:43 crc kubenswrapper[4904]: E0214 11:23:43.616491 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04" Feb 14 11:23:43 crc kubenswrapper[4904]: E0214 11:23:43.616675 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-82gcc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-68f46476f-4zj5b_openstack-operators(fa0069c3-fd8a-4ffe-aad0-ccee78f70caa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:23:43 crc kubenswrapper[4904]: E0214 11:23:43.618743 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-4zj5b" podUID="fa0069c3-fd8a-4ffe-aad0-ccee78f70caa" Feb 14 11:23:44 crc kubenswrapper[4904]: E0214 11:23:44.411632 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:e4689246ae78635dc3c1db9c677d8b16b8f94276df15fb9c84bfc57cc6578fcf" Feb 14 11:23:44 crc kubenswrapper[4904]: E0214 11:23:44.411827 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:e4689246ae78635dc3c1db9c677d8b16b8f94276df15fb9c84bfc57cc6578fcf,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4p5pm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-64ddbf8bb-hsx4c_openstack-operators(bac77172-b950-473c-8b07-0b1201a981db): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:23:44 crc kubenswrapper[4904]: E0214 11:23:44.413046 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-hsx4c" podUID="bac77172-b950-473c-8b07-0b1201a981db" Feb 14 11:23:44 crc kubenswrapper[4904]: E0214 11:23:44.575495 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-4zj5b" podUID="fa0069c3-fd8a-4ffe-aad0-ccee78f70caa" Feb 14 11:23:44 crc kubenswrapper[4904]: E0214 11:23:44.581694 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:e4689246ae78635dc3c1db9c677d8b16b8f94276df15fb9c84bfc57cc6578fcf\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-hsx4c" podUID="bac77172-b950-473c-8b07-0b1201a981db" Feb 14 11:23:45 crc kubenswrapper[4904]: E0214 11:23:45.757657 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9f2e1299d908411457e53b49e1062265d2b9d76f6719db24d1be9347c388e4da" Feb 14 11:23:45 crc kubenswrapper[4904]: E0214 11:23:45.757849 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9f2e1299d908411457e53b49e1062265d2b9d76f6719db24d1be9347c388e4da,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sbjqp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-5b9b8895d5-m8p4r_openstack-operators(d4632ec1-624e-4097-9888-0bbe395d6916): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:23:45 crc kubenswrapper[4904]: E0214 11:23:45.759044 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-m8p4r" podUID="d4632ec1-624e-4097-9888-0bbe395d6916" Feb 14 11:23:46 crc kubenswrapper[4904]: E0214 11:23:46.395192 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:a18f12497b7159b100fcfd72c7ba2273d0669a5c00600a9ff1333bca028f256a" Feb 14 11:23:46 crc kubenswrapper[4904]: E0214 11:23:46.395658 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:a18f12497b7159b100fcfd72c7ba2273d0669a5c00600a9ff1333bca028f256a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-66fcr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-6994f66f48-rvcg7_openstack-operators(b80a5769-6d09-4697-ba97-a5723e161a3c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:23:46 crc kubenswrapper[4904]: E0214 11:23:46.396880 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-rvcg7" podUID="b80a5769-6d09-4697-ba97-a5723e161a3c" Feb 14 11:23:46 crc kubenswrapper[4904]: E0214 11:23:46.587557 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:a18f12497b7159b100fcfd72c7ba2273d0669a5c00600a9ff1333bca028f256a\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-rvcg7" podUID="b80a5769-6d09-4697-ba97-a5723e161a3c" Feb 14 11:23:46 crc kubenswrapper[4904]: E0214 11:23:46.588857 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:9f2e1299d908411457e53b49e1062265d2b9d76f6719db24d1be9347c388e4da\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-m8p4r" podUID="d4632ec1-624e-4097-9888-0bbe395d6916" Feb 14 11:23:48 crc kubenswrapper[4904]: E0214 11:23:48.753315 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:8fb0a33b8d93cf9f84f079af5f2ceb680afada4e44542514959146779f57f64c" Feb 14 11:23:48 crc kubenswrapper[4904]: E0214 11:23:48.753496 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:8fb0a33b8d93cf9f84f079af5f2ceb680afada4e44542514959146779f57f64c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t4gfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-54f6768c69-6hjt5_openstack-operators(23733893-aa4e-47e2-a06a-a055313308d8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:23:48 crc kubenswrapper[4904]: E0214 11:23:48.754662 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-6hjt5" podUID="23733893-aa4e-47e2-a06a-a055313308d8" Feb 14 11:23:49 crc kubenswrapper[4904]: E0214 11:23:49.202335 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:1ab3ec59cd8e30dd8423e91ad832403bdefbae3b8ac47e15578d5a677d7ba0df" Feb 14 11:23:49 crc kubenswrapper[4904]: E0214 11:23:49.202581 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:1ab3ec59cd8e30dd8423e91ad832403bdefbae3b8ac47e15578d5a677d7ba0df,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9twzr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987464f4-jsn2d_openstack-operators(2db31100-df1f-441c-87bf-944dc247a4fc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:23:49 crc kubenswrapper[4904]: E0214 11:23:49.203761 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-77987464f4-jsn2d" podUID="2db31100-df1f-441c-87bf-944dc247a4fc" Feb 14 11:23:49 crc kubenswrapper[4904]: E0214 11:23:49.609518 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:8fb0a33b8d93cf9f84f079af5f2ceb680afada4e44542514959146779f57f64c\\\"\"" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-6hjt5" podUID="23733893-aa4e-47e2-a06a-a055313308d8" Feb 14 11:23:49 crc kubenswrapper[4904]: E0214 11:23:49.610128 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:1ab3ec59cd8e30dd8423e91ad832403bdefbae3b8ac47e15578d5a677d7ba0df\\\"\"" pod="openstack-operators/glance-operator-controller-manager-77987464f4-jsn2d" podUID="2db31100-df1f-441c-87bf-944dc247a4fc" Feb 14 11:23:49 crc kubenswrapper[4904]: E0214 11:23:49.637419 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:2b8ab3063af4aaeed0198197aae6f391c6647ac686c94c85668537f1d5933979" Feb 14 11:23:49 crc kubenswrapper[4904]: E0214 11:23:49.637676 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:2b8ab3063af4aaeed0198197aae6f391c6647ac686c94c85668537f1d5933979,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x7c4r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-5d946d989d-hbqp9_openstack-operators(33670c4d-fd2a-49cd-b602-ccdd927b7c00): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:23:49 crc kubenswrapper[4904]: E0214 11:23:49.642047 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-hbqp9" podUID="33670c4d-fd2a-49cd-b602-ccdd927b7c00" Feb 14 11:23:50 crc kubenswrapper[4904]: E0214 11:23:50.614660 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:2b8ab3063af4aaeed0198197aae6f391c6647ac686c94c85668537f1d5933979\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-hbqp9" podUID="33670c4d-fd2a-49cd-b602-ccdd927b7c00" Feb 14 11:23:53 crc kubenswrapper[4904]: E0214 11:23:53.364284 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838" Feb 14 11:23:53 crc kubenswrapper[4904]: E0214 11:23:53.366984 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tgfkx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-567668f5cf-qd78d_openstack-operators(f29b166d-edb4-4403-be20-a523578848ee): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:23:53 crc kubenswrapper[4904]: E0214 11:23:53.368178 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-qd78d" podUID="f29b166d-edb4-4403-be20-a523578848ee" Feb 14 11:23:53 crc kubenswrapper[4904]: E0214 11:23:53.631567 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838\\\"\"" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-qd78d" podUID="f29b166d-edb4-4403-be20-a523578848ee" Feb 14 11:23:53 crc kubenswrapper[4904]: E0214 11:23:53.817342 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1" Feb 14 11:23:53 crc kubenswrapper[4904]: E0214 11:23:53.817548 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7wdn7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-b4d948c87-sjlh5_openstack-operators(56b8547b-7339-452b-9a06-a05411222f70): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:23:53 crc kubenswrapper[4904]: E0214 11:23:53.818907 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-sjlh5" podUID="56b8547b-7339-452b-9a06-a05411222f70" Feb 14 11:23:54 crc kubenswrapper[4904]: E0214 11:23:54.614598 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Feb 14 11:23:54 crc kubenswrapper[4904]: E0214 11:23:54.614753 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-skl4f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-795l4_openstack-operators(aeb77a69-e9d9-4e04-a270-665a28bb0fa1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:23:54 crc kubenswrapper[4904]: E0214 11:23:54.615925 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-795l4" podUID="aeb77a69-e9d9-4e04-a270-665a28bb0fa1" Feb 14 11:23:54 crc kubenswrapper[4904]: E0214 11:23:54.637189 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-sjlh5" podUID="56b8547b-7339-452b-9a06-a05411222f70" Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.167087 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685"] Feb 14 11:23:56 crc kubenswrapper[4904]: W0214 11:23:56.175245 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3d1040e_0300_49ce_8b7d_7e4dac11c874.slice/crio-cb3e29ec99e22c900138784fe132f9e066db48d5941dd442e5e6de31f71c391b WatchSource:0}: Error finding container cb3e29ec99e22c900138784fe132f9e066db48d5941dd442e5e6de31f71c391b: Status 404 returned error can't find the container with id cb3e29ec99e22c900138784fe132f9e066db48d5941dd442e5e6de31f71c391b Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.359825 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8"] Feb 14 11:23:56 crc kubenswrapper[4904]: W0214 11:23:56.360089 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b9716ac_90dd_48ff_9dc8_7e5ef53ce0e9.slice/crio-d51421a90a6d28d343114805976c181a22fa2cc05d9b38fc1afe0eaa86565650 WatchSource:0}: Error finding container d51421a90a6d28d343114805976c181a22fa2cc05d9b38fc1afe0eaa86565650: Status 404 returned error can't find the container with id d51421a90a6d28d343114805976c181a22fa2cc05d9b38fc1afe0eaa86565650 Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.647536 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-mbrvg" event={"ID":"96c0bf26-aa73-4948-add8-b30fbbc23df9","Type":"ContainerStarted","Data":"756e2e17ea1707b2d91bfef0477bf5db12f72d8cdca6ab00d9bdda819ef21218"} Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.648391 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-mbrvg" Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.649869 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-gt547" event={"ID":"6c57743e-c864-4e0d-828a-406978aab771","Type":"ContainerStarted","Data":"2e94f1d49fa1913122c857a9b55f71bae22126b86a73ec6321a5cafbe3dffad8"} Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.650207 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-gt547" Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.651727 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q5cmd" event={"ID":"95a68c48-4e90-43c0-9caf-d6be291fca1e","Type":"ContainerStarted","Data":"8b1a6bd684acb8cdd6482d900b90df90455dae9b39cdafe19cbff86f07fae29d"} Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.652140 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q5cmd" Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.653659 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685" event={"ID":"c3d1040e-0300-49ce-8b7d-7e4dac11c874","Type":"ContainerStarted","Data":"cb3e29ec99e22c900138784fe132f9e066db48d5941dd442e5e6de31f71c391b"} Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.655273 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x8k6m" event={"ID":"53ba6432-a3b1-4cce-8ab2-bc74a718b134","Type":"ContainerStarted","Data":"a50b962b14ca737d84048f836b14018134d5fca9f318f40f451caade819edb09"} Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.655578 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x8k6m" Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.656628 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8" event={"ID":"8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9","Type":"ContainerStarted","Data":"d51421a90a6d28d343114805976c181a22fa2cc05d9b38fc1afe0eaa86565650"} Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.658114 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-k28jp" event={"ID":"8510a96b-51e8-4ad4-be1e-0fa7e2a86880","Type":"ContainerStarted","Data":"01803f1c0e25e3f50e4d55123df6c1ef4d94296fbe21aecc61e5842076dd5941"} Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.658255 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-k28jp" Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.659591 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7866795846-jnd94" event={"ID":"00d1bd7d-c3d8-4a54-9208-c28355796751","Type":"ContainerStarted","Data":"58ea2708ff83ec01f54e132f229c1f9db1267441647e504f457ef6c09189d12d"} Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.659779 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-7866795846-jnd94" Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.660961 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-xb6dd" event={"ID":"cebffdc5-38fd-45a5-ab0a-7445696c1661","Type":"ContainerStarted","Data":"c582e5dc9b70eeb6c5301fcead04bbc4b794541abc2b727181e3cef8d01cc55c"} Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.661556 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-xb6dd" Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.662968 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-pwsvp" event={"ID":"70a8fe80-89cb-492b-a631-e8042e3ed1be","Type":"ContainerStarted","Data":"4149464ffb91494faffbdb57173b422376a21820cb607ef67fbf1287bbd8c1e2"} Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.663435 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-pwsvp" Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.664881 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-m29fl" event={"ID":"2b26d95c-682d-4be5-9788-5044e186611c","Type":"ContainerStarted","Data":"98eb2586bfeffe4a4d2d31a3fe190dcd2e38aee9f47c4c6e6941f54251c0f461"} Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.665283 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-m29fl" Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.719672 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-mbrvg" podStartSLOduration=3.818982903 podStartE2EDuration="31.719653157s" podCreationTimestamp="2026-02-14 11:23:25 +0000 UTC" firstStartedPulling="2026-02-14 11:23:27.745155642 +0000 UTC m=+798.557920303" lastFinishedPulling="2026-02-14 11:23:55.645825886 +0000 UTC m=+826.458590557" observedRunningTime="2026-02-14 11:23:56.693064553 +0000 UTC m=+827.505829214" watchObservedRunningTime="2026-02-14 11:23:56.719653157 +0000 UTC m=+827.532417808" Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.745608 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-gt547" podStartSLOduration=4.152554081 podStartE2EDuration="31.745588813s" podCreationTimestamp="2026-02-14 11:23:25 +0000 UTC" firstStartedPulling="2026-02-14 11:23:27.531021611 +0000 UTC m=+798.343786262" lastFinishedPulling="2026-02-14 11:23:55.124056333 +0000 UTC m=+825.936820994" observedRunningTime="2026-02-14 11:23:56.741112099 +0000 UTC m=+827.553876760" watchObservedRunningTime="2026-02-14 11:23:56.745588813 +0000 UTC m=+827.558353474" Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.804069 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q5cmd" podStartSLOduration=3.713509193 podStartE2EDuration="30.804053336s" podCreationTimestamp="2026-02-14 11:23:26 +0000 UTC" firstStartedPulling="2026-02-14 11:23:28.632519405 +0000 UTC m=+799.445284066" lastFinishedPulling="2026-02-14 11:23:55.723063548 +0000 UTC m=+826.535828209" observedRunningTime="2026-02-14 11:23:56.797596778 +0000 UTC m=+827.610361439" watchObservedRunningTime="2026-02-14 11:23:56.804053336 +0000 UTC m=+827.616818007" Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.848868 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-m29fl" podStartSLOduration=5.329364763 podStartE2EDuration="31.848846743s" podCreationTimestamp="2026-02-14 11:23:25 +0000 UTC" firstStartedPulling="2026-02-14 11:23:27.300782455 +0000 UTC m=+798.113547116" lastFinishedPulling="2026-02-14 11:23:53.820264435 +0000 UTC m=+824.633029096" observedRunningTime="2026-02-14 11:23:56.84405413 +0000 UTC m=+827.656818791" watchObservedRunningTime="2026-02-14 11:23:56.848846743 +0000 UTC m=+827.661611414" Feb 14 11:23:56 crc kubenswrapper[4904]: I0214 11:23:56.955267 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-xb6dd" podStartSLOduration=3.754455895 podStartE2EDuration="30.95525265s" podCreationTimestamp="2026-02-14 11:23:26 +0000 UTC" firstStartedPulling="2026-02-14 11:23:28.444921078 +0000 UTC m=+799.257685739" lastFinishedPulling="2026-02-14 11:23:55.645717833 +0000 UTC m=+826.458482494" observedRunningTime="2026-02-14 11:23:56.945396898 +0000 UTC m=+827.758161559" watchObservedRunningTime="2026-02-14 11:23:56.95525265 +0000 UTC m=+827.768017311" Feb 14 11:23:57 crc kubenswrapper[4904]: I0214 11:23:57.017343 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x8k6m" podStartSLOduration=3.945073217 podStartE2EDuration="31.017328183s" podCreationTimestamp="2026-02-14 11:23:26 +0000 UTC" firstStartedPulling="2026-02-14 11:23:28.633459722 +0000 UTC m=+799.446224373" lastFinishedPulling="2026-02-14 11:23:55.705714678 +0000 UTC m=+826.518479339" observedRunningTime="2026-02-14 11:23:57.015589106 +0000 UTC m=+827.828353767" watchObservedRunningTime="2026-02-14 11:23:57.017328183 +0000 UTC m=+827.830092844" Feb 14 11:23:57 crc kubenswrapper[4904]: I0214 11:23:57.068900 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-pwsvp" podStartSLOduration=3.868294897 podStartE2EDuration="31.068884527s" podCreationTimestamp="2026-02-14 11:23:26 +0000 UTC" firstStartedPulling="2026-02-14 11:23:28.445194855 +0000 UTC m=+799.257959506" lastFinishedPulling="2026-02-14 11:23:55.645784475 +0000 UTC m=+826.458549136" observedRunningTime="2026-02-14 11:23:57.051419015 +0000 UTC m=+827.864183676" watchObservedRunningTime="2026-02-14 11:23:57.068884527 +0000 UTC m=+827.881649188" Feb 14 11:23:57 crc kubenswrapper[4904]: I0214 11:23:57.093729 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-7866795846-jnd94" podStartSLOduration=3.98613483 podStartE2EDuration="31.093709862s" podCreationTimestamp="2026-02-14 11:23:26 +0000 UTC" firstStartedPulling="2026-02-14 11:23:28.614377355 +0000 UTC m=+799.427142016" lastFinishedPulling="2026-02-14 11:23:55.721952377 +0000 UTC m=+826.534717048" observedRunningTime="2026-02-14 11:23:57.089860666 +0000 UTC m=+827.902625327" watchObservedRunningTime="2026-02-14 11:23:57.093709862 +0000 UTC m=+827.906474533" Feb 14 11:23:57 crc kubenswrapper[4904]: I0214 11:23:57.123687 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-k28jp" podStartSLOduration=4.430247835 podStartE2EDuration="32.123670639s" podCreationTimestamp="2026-02-14 11:23:25 +0000 UTC" firstStartedPulling="2026-02-14 11:23:27.951674742 +0000 UTC m=+798.764439403" lastFinishedPulling="2026-02-14 11:23:55.645097546 +0000 UTC m=+826.457862207" observedRunningTime="2026-02-14 11:23:57.122208198 +0000 UTC m=+827.934972879" watchObservedRunningTime="2026-02-14 11:23:57.123670639 +0000 UTC m=+827.936435300" Feb 14 11:23:58 crc kubenswrapper[4904]: I0214 11:23:58.693072 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-9gwm7" event={"ID":"733ceff2-fece-49be-8bc4-63f3adcfff22","Type":"ContainerStarted","Data":"b35d9075d24a091a8d89ee45fa6012a1322f5fee99d5e8e0cad692e93fa3be8c"} Feb 14 11:23:58 crc kubenswrapper[4904]: I0214 11:23:58.700205 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-4zj5b" event={"ID":"fa0069c3-fd8a-4ffe-aad0-ccee78f70caa","Type":"ContainerStarted","Data":"7b27d84a0e0a84cad93c08b6a9ff742ad231489357ca4b4e4c8627d4d9b50b82"} Feb 14 11:23:58 crc kubenswrapper[4904]: I0214 11:23:58.718555 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-68f46476f-4zj5b" podStartSLOduration=4.8883160530000005 podStartE2EDuration="32.718540122s" podCreationTimestamp="2026-02-14 11:23:26 +0000 UTC" firstStartedPulling="2026-02-14 11:23:28.554222555 +0000 UTC m=+799.366987226" lastFinishedPulling="2026-02-14 11:23:56.384446634 +0000 UTC m=+827.197211295" observedRunningTime="2026-02-14 11:23:58.714878892 +0000 UTC m=+829.527643563" watchObservedRunningTime="2026-02-14 11:23:58.718540122 +0000 UTC m=+829.531304783" Feb 14 11:23:58 crc kubenswrapper[4904]: I0214 11:23:58.952729 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:58 crc kubenswrapper[4904]: I0214 11:23:58.952802 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:58 crc kubenswrapper[4904]: I0214 11:23:58.960070 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-metrics-certs\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:58 crc kubenswrapper[4904]: I0214 11:23:58.960381 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/850dfcee-db10-493e-8043-5e167cfe3a54-webhook-certs\") pod \"openstack-operator-controller-manager-67d96975f-hzkn7\" (UID: \"850dfcee-db10-493e-8043-5e167cfe3a54\") " pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:58 crc kubenswrapper[4904]: I0214 11:23:58.990129 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:23:59 crc kubenswrapper[4904]: I0214 11:23:59.251797 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7"] Feb 14 11:23:59 crc kubenswrapper[4904]: I0214 11:23:59.706608 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-9gwm7" Feb 14 11:23:59 crc kubenswrapper[4904]: I0214 11:23:59.731299 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-9gwm7" podStartSLOduration=5.775898344 podStartE2EDuration="33.731279448s" podCreationTimestamp="2026-02-14 11:23:26 +0000 UTC" firstStartedPulling="2026-02-14 11:23:28.428867604 +0000 UTC m=+799.241632265" lastFinishedPulling="2026-02-14 11:23:56.384248708 +0000 UTC m=+827.197013369" observedRunningTime="2026-02-14 11:23:59.723736609 +0000 UTC m=+830.536501290" watchObservedRunningTime="2026-02-14 11:23:59.731279448 +0000 UTC m=+830.544044109" Feb 14 11:24:00 crc kubenswrapper[4904]: I0214 11:24:00.722967 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-hsx4c" event={"ID":"bac77172-b950-473c-8b07-0b1201a981db","Type":"ContainerStarted","Data":"c09e8ed371cbf8ea36fd36c75d63b20fa2a2b25f31987843fddc403c35639953"} Feb 14 11:24:00 crc kubenswrapper[4904]: I0214 11:24:00.723707 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-hsx4c" Feb 14 11:24:00 crc kubenswrapper[4904]: I0214 11:24:00.741200 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-hsx4c" podStartSLOduration=3.776898005 podStartE2EDuration="34.741183364s" podCreationTimestamp="2026-02-14 11:23:26 +0000 UTC" firstStartedPulling="2026-02-14 11:23:28.193341823 +0000 UTC m=+799.006125295" lastFinishedPulling="2026-02-14 11:23:59.157645993 +0000 UTC m=+829.970410654" observedRunningTime="2026-02-14 11:24:00.738652564 +0000 UTC m=+831.551417225" watchObservedRunningTime="2026-02-14 11:24:00.741183364 +0000 UTC m=+831.553948025" Feb 14 11:24:01 crc kubenswrapper[4904]: I0214 11:24:01.733850 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987464f4-jsn2d" event={"ID":"2db31100-df1f-441c-87bf-944dc247a4fc","Type":"ContainerStarted","Data":"c88560926a41601f387e7d64358af006a11634cb1b9119552d734d68448abf74"} Feb 14 11:24:01 crc kubenswrapper[4904]: I0214 11:24:01.734299 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987464f4-jsn2d" Feb 14 11:24:01 crc kubenswrapper[4904]: I0214 11:24:01.735481 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8" event={"ID":"8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9","Type":"ContainerStarted","Data":"d876e31ffd0dfafcf264893a69bb149d937d67fad884cb99717bf4fe01b9706f"} Feb 14 11:24:01 crc kubenswrapper[4904]: I0214 11:24:01.735595 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8" Feb 14 11:24:01 crc kubenswrapper[4904]: I0214 11:24:01.737208 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" event={"ID":"850dfcee-db10-493e-8043-5e167cfe3a54","Type":"ContainerStarted","Data":"c36a374b8fd89f59a94021c9894908e1cb92e933c7da8cc725ad1d24f7341b96"} Feb 14 11:24:01 crc kubenswrapper[4904]: I0214 11:24:01.737234 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" event={"ID":"850dfcee-db10-493e-8043-5e167cfe3a54","Type":"ContainerStarted","Data":"47975991c306269fae0760aff1c930614556980b35917f4037521e6ec9a726fd"} Feb 14 11:24:01 crc kubenswrapper[4904]: I0214 11:24:01.737324 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:24:01 crc kubenswrapper[4904]: I0214 11:24:01.738414 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-m8p4r" event={"ID":"d4632ec1-624e-4097-9888-0bbe395d6916","Type":"ContainerStarted","Data":"616e499e99e3f817a90bc92811931c3bf9c6225a8f6e1ac5bf2cebb71cfe9b02"} Feb 14 11:24:01 crc kubenswrapper[4904]: I0214 11:24:01.738599 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-m8p4r" Feb 14 11:24:01 crc kubenswrapper[4904]: I0214 11:24:01.739977 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685" event={"ID":"c3d1040e-0300-49ce-8b7d-7e4dac11c874","Type":"ContainerStarted","Data":"fc113e6cf5f9dc66762c329f11d2540853553410daac18d8cbb9c215edd8e0c9"} Feb 14 11:24:01 crc kubenswrapper[4904]: I0214 11:24:01.740191 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685" Feb 14 11:24:01 crc kubenswrapper[4904]: I0214 11:24:01.804772 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685" podStartSLOduration=30.936461721 podStartE2EDuration="35.804756901s" podCreationTimestamp="2026-02-14 11:23:26 +0000 UTC" firstStartedPulling="2026-02-14 11:23:56.177335497 +0000 UTC m=+826.990100158" lastFinishedPulling="2026-02-14 11:24:01.045630657 +0000 UTC m=+831.858395338" observedRunningTime="2026-02-14 11:24:01.800323819 +0000 UTC m=+832.613088480" watchObservedRunningTime="2026-02-14 11:24:01.804756901 +0000 UTC m=+832.617521562" Feb 14 11:24:01 crc kubenswrapper[4904]: I0214 11:24:01.805660 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987464f4-jsn2d" podStartSLOduration=3.311282537 podStartE2EDuration="36.805655506s" podCreationTimestamp="2026-02-14 11:23:25 +0000 UTC" firstStartedPulling="2026-02-14 11:23:27.553001507 +0000 UTC m=+798.365766158" lastFinishedPulling="2026-02-14 11:24:01.047374466 +0000 UTC m=+831.860139127" observedRunningTime="2026-02-14 11:24:01.767125843 +0000 UTC m=+832.579890504" watchObservedRunningTime="2026-02-14 11:24:01.805655506 +0000 UTC m=+832.618420167" Feb 14 11:24:01 crc kubenswrapper[4904]: I0214 11:24:01.835916 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" podStartSLOduration=35.835903282 podStartE2EDuration="35.835903282s" podCreationTimestamp="2026-02-14 11:23:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:24:01.835366596 +0000 UTC m=+832.648131257" watchObservedRunningTime="2026-02-14 11:24:01.835903282 +0000 UTC m=+832.648667943" Feb 14 11:24:01 crc kubenswrapper[4904]: I0214 11:24:01.861239 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8" podStartSLOduration=32.183077159 podStartE2EDuration="36.86122428s" podCreationTimestamp="2026-02-14 11:23:25 +0000 UTC" firstStartedPulling="2026-02-14 11:23:56.378418618 +0000 UTC m=+827.191183279" lastFinishedPulling="2026-02-14 11:24:01.056565729 +0000 UTC m=+831.869330400" observedRunningTime="2026-02-14 11:24:01.856635794 +0000 UTC m=+832.669400465" watchObservedRunningTime="2026-02-14 11:24:01.86122428 +0000 UTC m=+832.673988941" Feb 14 11:24:01 crc kubenswrapper[4904]: I0214 11:24:01.876042 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-m8p4r" podStartSLOduration=3.442277744 podStartE2EDuration="36.876026648s" podCreationTimestamp="2026-02-14 11:23:25 +0000 UTC" firstStartedPulling="2026-02-14 11:23:27.622933548 +0000 UTC m=+798.435698209" lastFinishedPulling="2026-02-14 11:24:01.056682442 +0000 UTC m=+831.869447113" observedRunningTime="2026-02-14 11:24:01.873091487 +0000 UTC m=+832.685856148" watchObservedRunningTime="2026-02-14 11:24:01.876026648 +0000 UTC m=+832.688791309" Feb 14 11:24:02 crc kubenswrapper[4904]: I0214 11:24:02.747604 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-rvcg7" event={"ID":"b80a5769-6d09-4697-ba97-a5723e161a3c","Type":"ContainerStarted","Data":"de526178d2431f6c14966228f30b114ba50aa93c5938c87f419b434a66f56619"} Feb 14 11:24:02 crc kubenswrapper[4904]: I0214 11:24:02.766258 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-rvcg7" podStartSLOduration=2.715719743 podStartE2EDuration="36.766239101s" podCreationTimestamp="2026-02-14 11:23:26 +0000 UTC" firstStartedPulling="2026-02-14 11:23:28.188970443 +0000 UTC m=+799.001735104" lastFinishedPulling="2026-02-14 11:24:02.239489801 +0000 UTC m=+833.052254462" observedRunningTime="2026-02-14 11:24:02.760437401 +0000 UTC m=+833.573202062" watchObservedRunningTime="2026-02-14 11:24:02.766239101 +0000 UTC m=+833.579003752" Feb 14 11:24:03 crc kubenswrapper[4904]: I0214 11:24:03.754734 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-6hjt5" event={"ID":"23733893-aa4e-47e2-a06a-a055313308d8","Type":"ContainerStarted","Data":"071151cd490a7fb60b4a6b89a1b31393929a4739d87e2a0b8943c0043688effc"} Feb 14 11:24:03 crc kubenswrapper[4904]: I0214 11:24:03.755239 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-6hjt5" Feb 14 11:24:03 crc kubenswrapper[4904]: I0214 11:24:03.773405 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-6hjt5" podStartSLOduration=2.7064449550000003 podStartE2EDuration="37.773379781s" podCreationTimestamp="2026-02-14 11:23:26 +0000 UTC" firstStartedPulling="2026-02-14 11:23:28.220618006 +0000 UTC m=+799.033382667" lastFinishedPulling="2026-02-14 11:24:03.287552832 +0000 UTC m=+834.100317493" observedRunningTime="2026-02-14 11:24:03.767073428 +0000 UTC m=+834.579838109" watchObservedRunningTime="2026-02-14 11:24:03.773379781 +0000 UTC m=+834.586144452" Feb 14 11:24:06 crc kubenswrapper[4904]: I0214 11:24:06.085238 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-m29fl" Feb 14 11:24:06 crc kubenswrapper[4904]: I0214 11:24:06.110379 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-gt547" Feb 14 11:24:06 crc kubenswrapper[4904]: I0214 11:24:06.235328 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987464f4-jsn2d" Feb 14 11:24:06 crc kubenswrapper[4904]: I0214 11:24:06.334959 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-m8p4r" Feb 14 11:24:06 crc kubenswrapper[4904]: I0214 11:24:06.627729 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-k28jp" Feb 14 11:24:06 crc kubenswrapper[4904]: I0214 11:24:06.628407 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-mbrvg" Feb 14 11:24:06 crc kubenswrapper[4904]: I0214 11:24:06.750564 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-hsx4c" Feb 14 11:24:06 crc kubenswrapper[4904]: I0214 11:24:06.777698 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-hbqp9" event={"ID":"33670c4d-fd2a-49cd-b602-ccdd927b7c00","Type":"ContainerStarted","Data":"65e017401a7a9ac32cf91a1c134ad484fffe6df1f1e845aee1c2b6f273ecf0d7"} Feb 14 11:24:06 crc kubenswrapper[4904]: I0214 11:24:06.778916 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-hbqp9" Feb 14 11:24:06 crc kubenswrapper[4904]: I0214 11:24:06.805496 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-hbqp9" podStartSLOduration=2.6461329879999997 podStartE2EDuration="41.805478807s" podCreationTimestamp="2026-02-14 11:23:25 +0000 UTC" firstStartedPulling="2026-02-14 11:23:27.364625768 +0000 UTC m=+798.177390429" lastFinishedPulling="2026-02-14 11:24:06.523971587 +0000 UTC m=+837.336736248" observedRunningTime="2026-02-14 11:24:06.800372026 +0000 UTC m=+837.613136687" watchObservedRunningTime="2026-02-14 11:24:06.805478807 +0000 UTC m=+837.618243468" Feb 14 11:24:06 crc kubenswrapper[4904]: I0214 11:24:06.813855 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-rvcg7" Feb 14 11:24:06 crc kubenswrapper[4904]: I0214 11:24:06.833642 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-xb6dd" Feb 14 11:24:06 crc kubenswrapper[4904]: I0214 11:24:06.900154 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-9gwm7" Feb 14 11:24:06 crc kubenswrapper[4904]: I0214 11:24:06.939418 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-pwsvp" Feb 14 11:24:07 crc kubenswrapper[4904]: I0214 11:24:07.018350 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-68f46476f-4zj5b" Feb 14 11:24:07 crc kubenswrapper[4904]: I0214 11:24:07.020941 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-68f46476f-4zj5b" Feb 14 11:24:07 crc kubenswrapper[4904]: I0214 11:24:07.104898 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-q5cmd" Feb 14 11:24:07 crc kubenswrapper[4904]: I0214 11:24:07.178312 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-7866795846-jnd94" Feb 14 11:24:07 crc kubenswrapper[4904]: I0214 11:24:07.222507 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-x8k6m" Feb 14 11:24:07 crc kubenswrapper[4904]: I0214 11:24:07.784925 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-sjlh5" event={"ID":"56b8547b-7339-452b-9a06-a05411222f70","Type":"ContainerStarted","Data":"70803af929890499257c2bd64e20a50ebbdaef576c412ebfd0f60ec565cf89a1"} Feb 14 11:24:07 crc kubenswrapper[4904]: I0214 11:24:07.786375 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-sjlh5" Feb 14 11:24:07 crc kubenswrapper[4904]: I0214 11:24:07.788419 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-qd78d" event={"ID":"f29b166d-edb4-4403-be20-a523578848ee","Type":"ContainerStarted","Data":"9a5ee6629b465ef68609d6991ea1f54677f243ccbdb65e50f3d5cf2ffb526eb6"} Feb 14 11:24:07 crc kubenswrapper[4904]: I0214 11:24:07.810587 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-sjlh5" podStartSLOduration=2.429225994 podStartE2EDuration="41.810567781s" podCreationTimestamp="2026-02-14 11:23:26 +0000 UTC" firstStartedPulling="2026-02-14 11:23:27.932218405 +0000 UTC m=+798.744983066" lastFinishedPulling="2026-02-14 11:24:07.313560192 +0000 UTC m=+838.126324853" observedRunningTime="2026-02-14 11:24:07.807105805 +0000 UTC m=+838.619870466" watchObservedRunningTime="2026-02-14 11:24:07.810567781 +0000 UTC m=+838.623332442" Feb 14 11:24:08 crc kubenswrapper[4904]: I0214 11:24:08.996753 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-67d96975f-hzkn7" Feb 14 11:24:09 crc kubenswrapper[4904]: I0214 11:24:09.024013 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-qd78d" podStartSLOduration=3.683340383 podStartE2EDuration="43.023995476s" podCreationTimestamp="2026-02-14 11:23:26 +0000 UTC" firstStartedPulling="2026-02-14 11:23:27.924677227 +0000 UTC m=+798.737441888" lastFinishedPulling="2026-02-14 11:24:07.26533232 +0000 UTC m=+838.078096981" observedRunningTime="2026-02-14 11:24:07.83156183 +0000 UTC m=+838.644326491" watchObservedRunningTime="2026-02-14 11:24:09.023995476 +0000 UTC m=+839.836760127" Feb 14 11:24:09 crc kubenswrapper[4904]: E0214 11:24:09.838277 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-795l4" podUID="aeb77a69-e9d9-4e04-a270-665a28bb0fa1" Feb 14 11:24:12 crc kubenswrapper[4904]: I0214 11:24:12.138701 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5478cb7c68-h92m8" Feb 14 11:24:12 crc kubenswrapper[4904]: I0214 11:24:12.485397 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84966cf5c46s685" Feb 14 11:24:16 crc kubenswrapper[4904]: I0214 11:24:16.102272 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-hbqp9" Feb 14 11:24:16 crc kubenswrapper[4904]: I0214 11:24:16.656582 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-sjlh5" Feb 14 11:24:16 crc kubenswrapper[4904]: I0214 11:24:16.679003 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-6hjt5" Feb 14 11:24:16 crc kubenswrapper[4904]: I0214 11:24:16.781491 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-qd78d" Feb 14 11:24:16 crc kubenswrapper[4904]: I0214 11:24:16.784327 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-qd78d" Feb 14 11:24:16 crc kubenswrapper[4904]: I0214 11:24:16.818084 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-rvcg7" Feb 14 11:24:24 crc kubenswrapper[4904]: I0214 11:24:24.837660 4904 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 14 11:24:25 crc kubenswrapper[4904]: I0214 11:24:25.912373 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-795l4" event={"ID":"aeb77a69-e9d9-4e04-a270-665a28bb0fa1","Type":"ContainerStarted","Data":"9b025115774699b4756c098ee04f23bdec0f6726add482ea49986d16e5fd40ec"} Feb 14 11:24:25 crc kubenswrapper[4904]: I0214 11:24:25.927593 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-795l4" podStartSLOduration=3.090391725 podStartE2EDuration="59.927572107s" podCreationTimestamp="2026-02-14 11:23:26 +0000 UTC" firstStartedPulling="2026-02-14 11:23:28.581700323 +0000 UTC m=+799.394464984" lastFinishedPulling="2026-02-14 11:24:25.418880705 +0000 UTC m=+856.231645366" observedRunningTime="2026-02-14 11:24:25.925262744 +0000 UTC m=+856.738027405" watchObservedRunningTime="2026-02-14 11:24:25.927572107 +0000 UTC m=+856.740336768" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.299034 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-g77fw"] Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.301022 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-g77fw" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.303322 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-6chqc" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.305291 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.306322 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.307330 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-g77fw"] Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.308929 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.412058 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-kszpx"] Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.413711 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-kszpx" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.419014 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.440488 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-kszpx"] Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.484444 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fc7a6f3-f356-4161-8860-13e399d395f1-config\") pod \"dnsmasq-dns-675f4bcbfc-g77fw\" (UID: \"1fc7a6f3-f356-4161-8860-13e399d395f1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-g77fw" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.484508 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6szmk\" (UniqueName: \"kubernetes.io/projected/1fc7a6f3-f356-4161-8860-13e399d395f1-kube-api-access-6szmk\") pod \"dnsmasq-dns-675f4bcbfc-g77fw\" (UID: \"1fc7a6f3-f356-4161-8860-13e399d395f1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-g77fw" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.586445 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp9f7\" (UniqueName: \"kubernetes.io/projected/0688c0d9-6ac3-4c0e-a3d7-7745cd618c99-kube-api-access-wp9f7\") pod \"dnsmasq-dns-78dd6ddcc-kszpx\" (UID: \"0688c0d9-6ac3-4c0e-a3d7-7745cd618c99\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kszpx" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.586555 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fc7a6f3-f356-4161-8860-13e399d395f1-config\") pod \"dnsmasq-dns-675f4bcbfc-g77fw\" (UID: \"1fc7a6f3-f356-4161-8860-13e399d395f1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-g77fw" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.586616 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6szmk\" (UniqueName: \"kubernetes.io/projected/1fc7a6f3-f356-4161-8860-13e399d395f1-kube-api-access-6szmk\") pod \"dnsmasq-dns-675f4bcbfc-g77fw\" (UID: \"1fc7a6f3-f356-4161-8860-13e399d395f1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-g77fw" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.586654 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0688c0d9-6ac3-4c0e-a3d7-7745cd618c99-config\") pod \"dnsmasq-dns-78dd6ddcc-kszpx\" (UID: \"0688c0d9-6ac3-4c0e-a3d7-7745cd618c99\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kszpx" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.586683 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0688c0d9-6ac3-4c0e-a3d7-7745cd618c99-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-kszpx\" (UID: \"0688c0d9-6ac3-4c0e-a3d7-7745cd618c99\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kszpx" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.587867 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fc7a6f3-f356-4161-8860-13e399d395f1-config\") pod \"dnsmasq-dns-675f4bcbfc-g77fw\" (UID: \"1fc7a6f3-f356-4161-8860-13e399d395f1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-g77fw" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.613955 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6szmk\" (UniqueName: \"kubernetes.io/projected/1fc7a6f3-f356-4161-8860-13e399d395f1-kube-api-access-6szmk\") pod \"dnsmasq-dns-675f4bcbfc-g77fw\" (UID: \"1fc7a6f3-f356-4161-8860-13e399d395f1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-g77fw" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.623464 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-g77fw" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.688120 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0688c0d9-6ac3-4c0e-a3d7-7745cd618c99-config\") pod \"dnsmasq-dns-78dd6ddcc-kszpx\" (UID: \"0688c0d9-6ac3-4c0e-a3d7-7745cd618c99\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kszpx" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.688496 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0688c0d9-6ac3-4c0e-a3d7-7745cd618c99-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-kszpx\" (UID: \"0688c0d9-6ac3-4c0e-a3d7-7745cd618c99\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kszpx" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.688540 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp9f7\" (UniqueName: \"kubernetes.io/projected/0688c0d9-6ac3-4c0e-a3d7-7745cd618c99-kube-api-access-wp9f7\") pod \"dnsmasq-dns-78dd6ddcc-kszpx\" (UID: \"0688c0d9-6ac3-4c0e-a3d7-7745cd618c99\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kszpx" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.689420 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0688c0d9-6ac3-4c0e-a3d7-7745cd618c99-config\") pod \"dnsmasq-dns-78dd6ddcc-kszpx\" (UID: \"0688c0d9-6ac3-4c0e-a3d7-7745cd618c99\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kszpx" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.689508 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0688c0d9-6ac3-4c0e-a3d7-7745cd618c99-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-kszpx\" (UID: \"0688c0d9-6ac3-4c0e-a3d7-7745cd618c99\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kszpx" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.711925 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp9f7\" (UniqueName: \"kubernetes.io/projected/0688c0d9-6ac3-4c0e-a3d7-7745cd618c99-kube-api-access-wp9f7\") pod \"dnsmasq-dns-78dd6ddcc-kszpx\" (UID: \"0688c0d9-6ac3-4c0e-a3d7-7745cd618c99\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kszpx" Feb 14 11:24:42 crc kubenswrapper[4904]: I0214 11:24:42.735616 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-kszpx" Feb 14 11:24:43 crc kubenswrapper[4904]: I0214 11:24:43.089971 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-g77fw"] Feb 14 11:24:43 crc kubenswrapper[4904]: W0214 11:24:43.101257 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1fc7a6f3_f356_4161_8860_13e399d395f1.slice/crio-fe0f60de7db8ec2afda7633117db4990d5f80f2c3cc1695b9f30ae68168c191b WatchSource:0}: Error finding container fe0f60de7db8ec2afda7633117db4990d5f80f2c3cc1695b9f30ae68168c191b: Status 404 returned error can't find the container with id fe0f60de7db8ec2afda7633117db4990d5f80f2c3cc1695b9f30ae68168c191b Feb 14 11:24:43 crc kubenswrapper[4904]: I0214 11:24:43.197024 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-kszpx"] Feb 14 11:24:43 crc kubenswrapper[4904]: W0214 11:24:43.198302 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0688c0d9_6ac3_4c0e_a3d7_7745cd618c99.slice/crio-a2f39eca9c6beaba2eb3b1a07b8d9f1185a54f4ef3fd290a6d07b31ce8ea7df3 WatchSource:0}: Error finding container a2f39eca9c6beaba2eb3b1a07b8d9f1185a54f4ef3fd290a6d07b31ce8ea7df3: Status 404 returned error can't find the container with id a2f39eca9c6beaba2eb3b1a07b8d9f1185a54f4ef3fd290a6d07b31ce8ea7df3 Feb 14 11:24:44 crc kubenswrapper[4904]: I0214 11:24:44.029999 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-g77fw" event={"ID":"1fc7a6f3-f356-4161-8860-13e399d395f1","Type":"ContainerStarted","Data":"fe0f60de7db8ec2afda7633117db4990d5f80f2c3cc1695b9f30ae68168c191b"} Feb 14 11:24:44 crc kubenswrapper[4904]: I0214 11:24:44.031726 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-kszpx" event={"ID":"0688c0d9-6ac3-4c0e-a3d7-7745cd618c99","Type":"ContainerStarted","Data":"a2f39eca9c6beaba2eb3b1a07b8d9f1185a54f4ef3fd290a6d07b31ce8ea7df3"} Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.191507 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-g77fw"] Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.236369 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qhdgb"] Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.237567 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.247187 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qhdgb"] Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.437402 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-config\") pod \"dnsmasq-dns-666b6646f7-qhdgb\" (UID: \"0e9b3f7a-913c-43cd-bf74-a5b4df81567c\") " pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.437454 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-dns-svc\") pod \"dnsmasq-dns-666b6646f7-qhdgb\" (UID: \"0e9b3f7a-913c-43cd-bf74-a5b4df81567c\") " pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.437475 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v99fl\" (UniqueName: \"kubernetes.io/projected/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-kube-api-access-v99fl\") pod \"dnsmasq-dns-666b6646f7-qhdgb\" (UID: \"0e9b3f7a-913c-43cd-bf74-a5b4df81567c\") " pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.511202 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-kszpx"] Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.540251 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-gwbsl"] Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.540761 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-dns-svc\") pod \"dnsmasq-dns-666b6646f7-qhdgb\" (UID: \"0e9b3f7a-913c-43cd-bf74-a5b4df81567c\") " pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.541428 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.544954 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-dns-svc\") pod \"dnsmasq-dns-666b6646f7-qhdgb\" (UID: \"0e9b3f7a-913c-43cd-bf74-a5b4df81567c\") " pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.545029 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v99fl\" (UniqueName: \"kubernetes.io/projected/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-kube-api-access-v99fl\") pod \"dnsmasq-dns-666b6646f7-qhdgb\" (UID: \"0e9b3f7a-913c-43cd-bf74-a5b4df81567c\") " pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.545485 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-config\") pod \"dnsmasq-dns-666b6646f7-qhdgb\" (UID: \"0e9b3f7a-913c-43cd-bf74-a5b4df81567c\") " pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.546139 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-config\") pod \"dnsmasq-dns-666b6646f7-qhdgb\" (UID: \"0e9b3f7a-913c-43cd-bf74-a5b4df81567c\") " pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.564874 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-gwbsl"] Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.616951 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v99fl\" (UniqueName: \"kubernetes.io/projected/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-kube-api-access-v99fl\") pod \"dnsmasq-dns-666b6646f7-qhdgb\" (UID: \"0e9b3f7a-913c-43cd-bf74-a5b4df81567c\") " pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.646161 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28504613-d82c-4903-a641-684ab0e0c4c3-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-gwbsl\" (UID: \"28504613-d82c-4903-a641-684ab0e0c4c3\") " pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.646225 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28504613-d82c-4903-a641-684ab0e0c4c3-config\") pod \"dnsmasq-dns-57d769cc4f-gwbsl\" (UID: \"28504613-d82c-4903-a641-684ab0e0c4c3\") " pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.646280 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhp84\" (UniqueName: \"kubernetes.io/projected/28504613-d82c-4903-a641-684ab0e0c4c3-kube-api-access-qhp84\") pod \"dnsmasq-dns-57d769cc4f-gwbsl\" (UID: \"28504613-d82c-4903-a641-684ab0e0c4c3\") " pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.747725 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28504613-d82c-4903-a641-684ab0e0c4c3-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-gwbsl\" (UID: \"28504613-d82c-4903-a641-684ab0e0c4c3\") " pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.748565 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28504613-d82c-4903-a641-684ab0e0c4c3-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-gwbsl\" (UID: \"28504613-d82c-4903-a641-684ab0e0c4c3\") " pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.748591 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28504613-d82c-4903-a641-684ab0e0c4c3-config\") pod \"dnsmasq-dns-57d769cc4f-gwbsl\" (UID: \"28504613-d82c-4903-a641-684ab0e0c4c3\") " pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.747816 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28504613-d82c-4903-a641-684ab0e0c4c3-config\") pod \"dnsmasq-dns-57d769cc4f-gwbsl\" (UID: \"28504613-d82c-4903-a641-684ab0e0c4c3\") " pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.748710 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhp84\" (UniqueName: \"kubernetes.io/projected/28504613-d82c-4903-a641-684ab0e0c4c3-kube-api-access-qhp84\") pod \"dnsmasq-dns-57d769cc4f-gwbsl\" (UID: \"28504613-d82c-4903-a641-684ab0e0c4c3\") " pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.772545 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhp84\" (UniqueName: \"kubernetes.io/projected/28504613-d82c-4903-a641-684ab0e0c4c3-kube-api-access-qhp84\") pod \"dnsmasq-dns-57d769cc4f-gwbsl\" (UID: \"28504613-d82c-4903-a641-684ab0e0c4c3\") " pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.854725 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" Feb 14 11:24:45 crc kubenswrapper[4904]: I0214 11:24:45.876866 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.363630 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.365050 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.379215 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.379392 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.379458 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.379746 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.379814 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.380080 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-9qlhw" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.380139 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.458888 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.460019 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2b06dd61-da9f-455f-b813-53300e4fbc10-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.460061 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ctwr\" (UniqueName: \"kubernetes.io/projected/2b06dd61-da9f-455f-b813-53300e4fbc10-kube-api-access-9ctwr\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.460099 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2b06dd61-da9f-455f-b813-53300e4fbc10-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.460138 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.460169 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.460187 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2b06dd61-da9f-455f-b813-53300e4fbc10-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.460214 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2b06dd61-da9f-455f-b813-53300e4fbc10-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.460243 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.460266 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.460293 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.460342 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2b06dd61-da9f-455f-b813-53300e4fbc10-config-data\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.535585 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qhdgb"] Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.562557 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2b06dd61-da9f-455f-b813-53300e4fbc10-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.562592 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.562627 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.562655 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.562691 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2b06dd61-da9f-455f-b813-53300e4fbc10-config-data\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.562711 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2b06dd61-da9f-455f-b813-53300e4fbc10-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.562726 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ctwr\" (UniqueName: \"kubernetes.io/projected/2b06dd61-da9f-455f-b813-53300e4fbc10-kube-api-access-9ctwr\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.562743 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2b06dd61-da9f-455f-b813-53300e4fbc10-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.562778 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.562793 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.562809 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2b06dd61-da9f-455f-b813-53300e4fbc10-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.564116 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2b06dd61-da9f-455f-b813-53300e4fbc10-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.565048 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2b06dd61-da9f-455f-b813-53300e4fbc10-config-data\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.565646 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2b06dd61-da9f-455f-b813-53300e4fbc10-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.565997 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.566713 4904 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.567081 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.589645 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.590155 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2b06dd61-da9f-455f-b813-53300e4fbc10-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.590530 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2b06dd61-da9f-455f-b813-53300e4fbc10-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.591866 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ctwr\" (UniqueName: \"kubernetes.io/projected/2b06dd61-da9f-455f-b813-53300e4fbc10-kube-api-access-9ctwr\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.592390 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.602512 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-gwbsl"] Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.631550 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.728316 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.754328 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.755858 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.765376 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.765613 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.765633 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.765882 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.766505 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.766538 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.769330 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-6jpz2" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.776251 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.868557 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgz54\" (UniqueName: \"kubernetes.io/projected/7448d80b-d340-46a2-88a6-221db1f93f11-kube-api-access-zgz54\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.868625 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.868669 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7448d80b-d340-46a2-88a6-221db1f93f11-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.868695 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.868725 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.868766 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7448d80b-d340-46a2-88a6-221db1f93f11-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.868797 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.868824 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7448d80b-d340-46a2-88a6-221db1f93f11-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.868875 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7448d80b-d340-46a2-88a6-221db1f93f11-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.868904 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.868935 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7448d80b-d340-46a2-88a6-221db1f93f11-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.971932 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.972004 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7448d80b-d340-46a2-88a6-221db1f93f11-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.972039 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7448d80b-d340-46a2-88a6-221db1f93f11-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.972093 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.972148 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7448d80b-d340-46a2-88a6-221db1f93f11-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.972274 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgz54\" (UniqueName: \"kubernetes.io/projected/7448d80b-d340-46a2-88a6-221db1f93f11-kube-api-access-zgz54\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.972334 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.972366 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7448d80b-d340-46a2-88a6-221db1f93f11-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.972416 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.972445 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.972505 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7448d80b-d340-46a2-88a6-221db1f93f11-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.974547 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7448d80b-d340-46a2-88a6-221db1f93f11-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.975290 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.976371 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.980144 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7448d80b-d340-46a2-88a6-221db1f93f11-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.981455 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.982405 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7448d80b-d340-46a2-88a6-221db1f93f11-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.984863 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.991975 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7448d80b-d340-46a2-88a6-221db1f93f11-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:46 crc kubenswrapper[4904]: I0214 11:24:46.999449 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7448d80b-d340-46a2-88a6-221db1f93f11-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:47 crc kubenswrapper[4904]: I0214 11:24:47.000599 4904 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:47 crc kubenswrapper[4904]: I0214 11:24:47.001065 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgz54\" (UniqueName: \"kubernetes.io/projected/7448d80b-d340-46a2-88a6-221db1f93f11-kube-api-access-zgz54\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:47 crc kubenswrapper[4904]: I0214 11:24:47.022202 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:47 crc kubenswrapper[4904]: I0214 11:24:47.078823 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" event={"ID":"0e9b3f7a-913c-43cd-bf74-a5b4df81567c","Type":"ContainerStarted","Data":"34a29d65190696d9cd41c56e03689ca50d1c4be9a071dc708aa3a0883caee803"} Feb 14 11:24:47 crc kubenswrapper[4904]: I0214 11:24:47.081401 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" event={"ID":"28504613-d82c-4903-a641-684ab0e0c4c3","Type":"ContainerStarted","Data":"8e1380ab4cd03cfd5371eb576851f1c94c2940feca272a5eb6a42c293c3a2abc"} Feb 14 11:24:47 crc kubenswrapper[4904]: I0214 11:24:47.132422 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:24:47 crc kubenswrapper[4904]: I0214 11:24:47.409950 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 14 11:24:47 crc kubenswrapper[4904]: I0214 11:24:47.834069 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.083312 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.084441 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.092954 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.094689 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-gqs8m" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.096344 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.097980 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.098669 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.109038 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2b06dd61-da9f-455f-b813-53300e4fbc10","Type":"ContainerStarted","Data":"f1209a24707701731254d181030bc253c4623d294926a0f93e47a725aa24a711"} Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.125581 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.206604 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-config-data-default\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.207390 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-kolla-config\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.207410 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.207480 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzkmk\" (UniqueName: \"kubernetes.io/projected/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-kube-api-access-vzkmk\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.207505 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.207528 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.207555 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.207584 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.308623 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-config-data-default\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.308669 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-kolla-config\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.308685 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.308710 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzkmk\" (UniqueName: \"kubernetes.io/projected/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-kube-api-access-vzkmk\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.308733 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.308751 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.308771 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.308798 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.309912 4904 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.310078 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.310376 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-kolla-config\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.310426 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-config-data-default\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.311107 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.315158 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.323767 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.327584 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.337726 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzkmk\" (UniqueName: \"kubernetes.io/projected/b6cb7af3-751c-48dd-84ae-6b77f7b9c233-kube-api-access-vzkmk\") pod \"openstack-galera-0\" (UID: \"b6cb7af3-751c-48dd-84ae-6b77f7b9c233\") " pod="openstack/openstack-galera-0" Feb 14 11:24:48 crc kubenswrapper[4904]: I0214 11:24:48.430491 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.248697 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.251924 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.252001 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.254694 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.257649 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.257967 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.258755 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-tdknq" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.331433 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c2878c0a-2515-4412-9aa7-acdb199bc463-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.331508 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.331543 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2878c0a-2515-4412-9aa7-acdb199bc463-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.331564 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2878c0a-2515-4412-9aa7-acdb199bc463-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.331586 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c2878c0a-2515-4412-9aa7-acdb199bc463-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.331603 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2878c0a-2515-4412-9aa7-acdb199bc463-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.331620 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c2878c0a-2515-4412-9aa7-acdb199bc463-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.331693 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gmgx\" (UniqueName: \"kubernetes.io/projected/c2878c0a-2515-4412-9aa7-acdb199bc463-kube-api-access-8gmgx\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.437011 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c2878c0a-2515-4412-9aa7-acdb199bc463-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.437126 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gmgx\" (UniqueName: \"kubernetes.io/projected/c2878c0a-2515-4412-9aa7-acdb199bc463-kube-api-access-8gmgx\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.437158 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c2878c0a-2515-4412-9aa7-acdb199bc463-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.437196 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.437223 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2878c0a-2515-4412-9aa7-acdb199bc463-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.437241 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2878c0a-2515-4412-9aa7-acdb199bc463-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.437260 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c2878c0a-2515-4412-9aa7-acdb199bc463-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.437276 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2878c0a-2515-4412-9aa7-acdb199bc463-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.439916 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2878c0a-2515-4412-9aa7-acdb199bc463-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.443526 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c2878c0a-2515-4412-9aa7-acdb199bc463-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.444378 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c2878c0a-2515-4412-9aa7-acdb199bc463-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.444766 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c2878c0a-2515-4412-9aa7-acdb199bc463-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.444888 4904 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.448965 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2878c0a-2515-4412-9aa7-acdb199bc463-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.450101 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2878c0a-2515-4412-9aa7-acdb199bc463-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.459951 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gmgx\" (UniqueName: \"kubernetes.io/projected/c2878c0a-2515-4412-9aa7-acdb199bc463-kube-api-access-8gmgx\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.508414 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c2878c0a-2515-4412-9aa7-acdb199bc463\") " pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.576448 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.606844 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.607871 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.611262 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.611298 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.611503 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-hpf6r" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.616926 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.641122 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a35e40d9-4ced-4411-aff4-68c295331d47-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a35e40d9-4ced-4411-aff4-68c295331d47\") " pod="openstack/memcached-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.641185 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a35e40d9-4ced-4411-aff4-68c295331d47-config-data\") pod \"memcached-0\" (UID: \"a35e40d9-4ced-4411-aff4-68c295331d47\") " pod="openstack/memcached-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.641228 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a35e40d9-4ced-4411-aff4-68c295331d47-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a35e40d9-4ced-4411-aff4-68c295331d47\") " pod="openstack/memcached-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.641253 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a35e40d9-4ced-4411-aff4-68c295331d47-kolla-config\") pod \"memcached-0\" (UID: \"a35e40d9-4ced-4411-aff4-68c295331d47\") " pod="openstack/memcached-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.641290 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzv2t\" (UniqueName: \"kubernetes.io/projected/a35e40d9-4ced-4411-aff4-68c295331d47-kube-api-access-qzv2t\") pod \"memcached-0\" (UID: \"a35e40d9-4ced-4411-aff4-68c295331d47\") " pod="openstack/memcached-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.742635 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a35e40d9-4ced-4411-aff4-68c295331d47-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a35e40d9-4ced-4411-aff4-68c295331d47\") " pod="openstack/memcached-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.742706 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a35e40d9-4ced-4411-aff4-68c295331d47-config-data\") pod \"memcached-0\" (UID: \"a35e40d9-4ced-4411-aff4-68c295331d47\") " pod="openstack/memcached-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.742761 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a35e40d9-4ced-4411-aff4-68c295331d47-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a35e40d9-4ced-4411-aff4-68c295331d47\") " pod="openstack/memcached-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.742795 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a35e40d9-4ced-4411-aff4-68c295331d47-kolla-config\") pod \"memcached-0\" (UID: \"a35e40d9-4ced-4411-aff4-68c295331d47\") " pod="openstack/memcached-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.742865 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzv2t\" (UniqueName: \"kubernetes.io/projected/a35e40d9-4ced-4411-aff4-68c295331d47-kube-api-access-qzv2t\") pod \"memcached-0\" (UID: \"a35e40d9-4ced-4411-aff4-68c295331d47\") " pod="openstack/memcached-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.744550 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a35e40d9-4ced-4411-aff4-68c295331d47-config-data\") pod \"memcached-0\" (UID: \"a35e40d9-4ced-4411-aff4-68c295331d47\") " pod="openstack/memcached-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.746020 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a35e40d9-4ced-4411-aff4-68c295331d47-kolla-config\") pod \"memcached-0\" (UID: \"a35e40d9-4ced-4411-aff4-68c295331d47\") " pod="openstack/memcached-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.765034 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a35e40d9-4ced-4411-aff4-68c295331d47-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a35e40d9-4ced-4411-aff4-68c295331d47\") " pod="openstack/memcached-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.765719 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a35e40d9-4ced-4411-aff4-68c295331d47-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a35e40d9-4ced-4411-aff4-68c295331d47\") " pod="openstack/memcached-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.768822 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzv2t\" (UniqueName: \"kubernetes.io/projected/a35e40d9-4ced-4411-aff4-68c295331d47-kube-api-access-qzv2t\") pod \"memcached-0\" (UID: \"a35e40d9-4ced-4411-aff4-68c295331d47\") " pod="openstack/memcached-0" Feb 14 11:24:49 crc kubenswrapper[4904]: I0214 11:24:49.951171 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 14 11:24:51 crc kubenswrapper[4904]: I0214 11:24:51.951060 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 14 11:24:51 crc kubenswrapper[4904]: I0214 11:24:51.952473 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 14 11:24:51 crc kubenswrapper[4904]: I0214 11:24:51.956296 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-gq9bv" Feb 14 11:24:51 crc kubenswrapper[4904]: I0214 11:24:51.978225 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 14 11:24:52 crc kubenswrapper[4904]: I0214 11:24:52.004808 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pssqp\" (UniqueName: \"kubernetes.io/projected/8515da2f-15c7-43af-b57f-bec14a6b0398-kube-api-access-pssqp\") pod \"kube-state-metrics-0\" (UID: \"8515da2f-15c7-43af-b57f-bec14a6b0398\") " pod="openstack/kube-state-metrics-0" Feb 14 11:24:52 crc kubenswrapper[4904]: I0214 11:24:52.106479 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pssqp\" (UniqueName: \"kubernetes.io/projected/8515da2f-15c7-43af-b57f-bec14a6b0398-kube-api-access-pssqp\") pod \"kube-state-metrics-0\" (UID: \"8515da2f-15c7-43af-b57f-bec14a6b0398\") " pod="openstack/kube-state-metrics-0" Feb 14 11:24:52 crc kubenswrapper[4904]: I0214 11:24:52.145197 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pssqp\" (UniqueName: \"kubernetes.io/projected/8515da2f-15c7-43af-b57f-bec14a6b0398-kube-api-access-pssqp\") pod \"kube-state-metrics-0\" (UID: \"8515da2f-15c7-43af-b57f-bec14a6b0398\") " pod="openstack/kube-state-metrics-0" Feb 14 11:24:52 crc kubenswrapper[4904]: I0214 11:24:52.274094 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.169490 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7448d80b-d340-46a2-88a6-221db1f93f11","Type":"ContainerStarted","Data":"1f9964be6c39625296f551e8e5b8f52617d69edd9c1bb0fdc725f9fdfe524226"} Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.526775 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zj52b"] Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.528259 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.531253 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-9cr7f" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.535669 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.536001 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.539704 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zj52b"] Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.625140 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-9jk9m"] Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.626533 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.637957 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-9jk9m"] Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.652578 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/05a401e2-caea-48a9-bb53-6a42b59dad57-scripts\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.652627 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/05a401e2-caea-48a9-bb53-6a42b59dad57-var-log-ovn\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.652647 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/05a401e2-caea-48a9-bb53-6a42b59dad57-ovn-controller-tls-certs\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.652689 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05a401e2-caea-48a9-bb53-6a42b59dad57-combined-ca-bundle\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.652710 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/05a401e2-caea-48a9-bb53-6a42b59dad57-var-run\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.652762 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/05a401e2-caea-48a9-bb53-6a42b59dad57-var-run-ovn\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.652791 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsj5x\" (UniqueName: \"kubernetes.io/projected/05a401e2-caea-48a9-bb53-6a42b59dad57-kube-api-access-nsj5x\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.754146 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/05a401e2-caea-48a9-bb53-6a42b59dad57-scripts\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.754191 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/05a401e2-caea-48a9-bb53-6a42b59dad57-var-log-ovn\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.754259 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/05a401e2-caea-48a9-bb53-6a42b59dad57-ovn-controller-tls-certs\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.754283 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0422124f-385d-4e36-9a12-29bd9a6fa72e-var-log\") pod \"ovn-controller-ovs-9jk9m\" (UID: \"0422124f-385d-4e36-9a12-29bd9a6fa72e\") " pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.754346 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05a401e2-caea-48a9-bb53-6a42b59dad57-combined-ca-bundle\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.754384 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0422124f-385d-4e36-9a12-29bd9a6fa72e-var-lib\") pod \"ovn-controller-ovs-9jk9m\" (UID: \"0422124f-385d-4e36-9a12-29bd9a6fa72e\") " pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.754464 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/05a401e2-caea-48a9-bb53-6a42b59dad57-var-run\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.755005 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/05a401e2-caea-48a9-bb53-6a42b59dad57-var-run\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.754507 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgr4f\" (UniqueName: \"kubernetes.io/projected/0422124f-385d-4e36-9a12-29bd9a6fa72e-kube-api-access-mgr4f\") pod \"ovn-controller-ovs-9jk9m\" (UID: \"0422124f-385d-4e36-9a12-29bd9a6fa72e\") " pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.756297 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/05a401e2-caea-48a9-bb53-6a42b59dad57-scripts\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.756322 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0422124f-385d-4e36-9a12-29bd9a6fa72e-etc-ovs\") pod \"ovn-controller-ovs-9jk9m\" (UID: \"0422124f-385d-4e36-9a12-29bd9a6fa72e\") " pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.756363 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0422124f-385d-4e36-9a12-29bd9a6fa72e-scripts\") pod \"ovn-controller-ovs-9jk9m\" (UID: \"0422124f-385d-4e36-9a12-29bd9a6fa72e\") " pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.756411 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/05a401e2-caea-48a9-bb53-6a42b59dad57-var-run-ovn\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.756437 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0422124f-385d-4e36-9a12-29bd9a6fa72e-var-run\") pod \"ovn-controller-ovs-9jk9m\" (UID: \"0422124f-385d-4e36-9a12-29bd9a6fa72e\") " pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.756489 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsj5x\" (UniqueName: \"kubernetes.io/projected/05a401e2-caea-48a9-bb53-6a42b59dad57-kube-api-access-nsj5x\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.757225 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/05a401e2-caea-48a9-bb53-6a42b59dad57-var-log-ovn\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.758960 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/05a401e2-caea-48a9-bb53-6a42b59dad57-var-run-ovn\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.760283 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/05a401e2-caea-48a9-bb53-6a42b59dad57-ovn-controller-tls-certs\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.775884 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05a401e2-caea-48a9-bb53-6a42b59dad57-combined-ca-bundle\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.780116 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsj5x\" (UniqueName: \"kubernetes.io/projected/05a401e2-caea-48a9-bb53-6a42b59dad57-kube-api-access-nsj5x\") pod \"ovn-controller-zj52b\" (UID: \"05a401e2-caea-48a9-bb53-6a42b59dad57\") " pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.858206 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0422124f-385d-4e36-9a12-29bd9a6fa72e-etc-ovs\") pod \"ovn-controller-ovs-9jk9m\" (UID: \"0422124f-385d-4e36-9a12-29bd9a6fa72e\") " pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.858257 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0422124f-385d-4e36-9a12-29bd9a6fa72e-scripts\") pod \"ovn-controller-ovs-9jk9m\" (UID: \"0422124f-385d-4e36-9a12-29bd9a6fa72e\") " pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.858282 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0422124f-385d-4e36-9a12-29bd9a6fa72e-var-run\") pod \"ovn-controller-ovs-9jk9m\" (UID: \"0422124f-385d-4e36-9a12-29bd9a6fa72e\") " pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.858333 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0422124f-385d-4e36-9a12-29bd9a6fa72e-var-log\") pod \"ovn-controller-ovs-9jk9m\" (UID: \"0422124f-385d-4e36-9a12-29bd9a6fa72e\") " pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.858372 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0422124f-385d-4e36-9a12-29bd9a6fa72e-var-lib\") pod \"ovn-controller-ovs-9jk9m\" (UID: \"0422124f-385d-4e36-9a12-29bd9a6fa72e\") " pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.858415 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgr4f\" (UniqueName: \"kubernetes.io/projected/0422124f-385d-4e36-9a12-29bd9a6fa72e-kube-api-access-mgr4f\") pod \"ovn-controller-ovs-9jk9m\" (UID: \"0422124f-385d-4e36-9a12-29bd9a6fa72e\") " pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.858446 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0422124f-385d-4e36-9a12-29bd9a6fa72e-var-run\") pod \"ovn-controller-ovs-9jk9m\" (UID: \"0422124f-385d-4e36-9a12-29bd9a6fa72e\") " pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.858497 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0422124f-385d-4e36-9a12-29bd9a6fa72e-etc-ovs\") pod \"ovn-controller-ovs-9jk9m\" (UID: \"0422124f-385d-4e36-9a12-29bd9a6fa72e\") " pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.858613 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0422124f-385d-4e36-9a12-29bd9a6fa72e-var-log\") pod \"ovn-controller-ovs-9jk9m\" (UID: \"0422124f-385d-4e36-9a12-29bd9a6fa72e\") " pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.859181 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0422124f-385d-4e36-9a12-29bd9a6fa72e-var-lib\") pod \"ovn-controller-ovs-9jk9m\" (UID: \"0422124f-385d-4e36-9a12-29bd9a6fa72e\") " pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.860148 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zj52b" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.861910 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0422124f-385d-4e36-9a12-29bd9a6fa72e-scripts\") pod \"ovn-controller-ovs-9jk9m\" (UID: \"0422124f-385d-4e36-9a12-29bd9a6fa72e\") " pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.874687 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgr4f\" (UniqueName: \"kubernetes.io/projected/0422124f-385d-4e36-9a12-29bd9a6fa72e-kube-api-access-mgr4f\") pod \"ovn-controller-ovs-9jk9m\" (UID: \"0422124f-385d-4e36-9a12-29bd9a6fa72e\") " pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:54 crc kubenswrapper[4904]: I0214 11:24:54.963572 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:24:55 crc kubenswrapper[4904]: I0214 11:24:55.989186 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 14 11:24:55 crc kubenswrapper[4904]: I0214 11:24:55.990854 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.008153 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.011657 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.011743 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-dzjrf" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.012283 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.012546 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.012746 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.082713 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/02df4ffa-81a6-4fd0-9136-3eedd479f653-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.082778 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02df4ffa-81a6-4fd0-9136-3eedd479f653-config\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.082906 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/02df4ffa-81a6-4fd0-9136-3eedd479f653-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.082937 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/02df4ffa-81a6-4fd0-9136-3eedd479f653-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.082976 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzf8z\" (UniqueName: \"kubernetes.io/projected/02df4ffa-81a6-4fd0-9136-3eedd479f653-kube-api-access-wzf8z\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.083018 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02df4ffa-81a6-4fd0-9136-3eedd479f653-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.083116 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/02df4ffa-81a6-4fd0-9136-3eedd479f653-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.083164 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.184315 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.184413 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/02df4ffa-81a6-4fd0-9136-3eedd479f653-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.184452 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02df4ffa-81a6-4fd0-9136-3eedd479f653-config\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.184489 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/02df4ffa-81a6-4fd0-9136-3eedd479f653-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.184512 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/02df4ffa-81a6-4fd0-9136-3eedd479f653-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.184549 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzf8z\" (UniqueName: \"kubernetes.io/projected/02df4ffa-81a6-4fd0-9136-3eedd479f653-kube-api-access-wzf8z\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.184586 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02df4ffa-81a6-4fd0-9136-3eedd479f653-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.184626 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/02df4ffa-81a6-4fd0-9136-3eedd479f653-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.186211 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/02df4ffa-81a6-4fd0-9136-3eedd479f653-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.186786 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/02df4ffa-81a6-4fd0-9136-3eedd479f653-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.186994 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02df4ffa-81a6-4fd0-9136-3eedd479f653-config\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.187084 4904 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.191764 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02df4ffa-81a6-4fd0-9136-3eedd479f653-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.193571 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/02df4ffa-81a6-4fd0-9136-3eedd479f653-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.198764 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/02df4ffa-81a6-4fd0-9136-3eedd479f653-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.202919 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzf8z\" (UniqueName: \"kubernetes.io/projected/02df4ffa-81a6-4fd0-9136-3eedd479f653-kube-api-access-wzf8z\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.212740 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"02df4ffa-81a6-4fd0-9136-3eedd479f653\") " pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:56 crc kubenswrapper[4904]: I0214 11:24:56.313789 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.490714 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.493315 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.502019 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.502328 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-x6gvm" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.502345 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.503773 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.504548 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.522446 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.522597 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-config\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.522625 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.522672 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.522726 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4vz5\" (UniqueName: \"kubernetes.io/projected/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-kube-api-access-g4vz5\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.522953 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.523005 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.523029 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.616624 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.626476 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.626523 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4vz5\" (UniqueName: \"kubernetes.io/projected/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-kube-api-access-g4vz5\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.626561 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.626627 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.626688 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.626738 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.626759 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-config\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.626775 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.627466 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.627538 4904 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.627955 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-config\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.628947 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.634279 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.634446 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.639927 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.642409 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4vz5\" (UniqueName: \"kubernetes.io/projected/fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c-kube-api-access-g4vz5\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.658623 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c\") " pod="openstack/ovsdbserver-sb-0" Feb 14 11:24:58 crc kubenswrapper[4904]: I0214 11:24:58.825014 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 14 11:25:04 crc kubenswrapper[4904]: W0214 11:25:04.617355 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8515da2f_15c7_43af_b57f_bec14a6b0398.slice/crio-c98bd77a5d4126fefa6d1970fdac9858507a41c462445f4c117339d578519602 WatchSource:0}: Error finding container c98bd77a5d4126fefa6d1970fdac9858507a41c462445f4c117339d578519602: Status 404 returned error can't find the container with id c98bd77a5d4126fefa6d1970fdac9858507a41c462445f4c117339d578519602 Feb 14 11:25:05 crc kubenswrapper[4904]: I0214 11:25:05.279626 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8515da2f-15c7-43af-b57f-bec14a6b0398","Type":"ContainerStarted","Data":"c98bd77a5d4126fefa6d1970fdac9858507a41c462445f4c117339d578519602"} Feb 14 11:25:05 crc kubenswrapper[4904]: E0214 11:25:05.463699 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 14 11:25:05 crc kubenswrapper[4904]: E0214 11:25:05.463974 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v99fl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-qhdgb_openstack(0e9b3f7a-913c-43cd-bf74-a5b4df81567c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:25:05 crc kubenswrapper[4904]: E0214 11:25:05.465196 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" podUID="0e9b3f7a-913c-43cd-bf74-a5b4df81567c" Feb 14 11:25:05 crc kubenswrapper[4904]: E0214 11:25:05.489678 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 14 11:25:05 crc kubenswrapper[4904]: E0214 11:25:05.489937 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6szmk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-g77fw_openstack(1fc7a6f3-f356-4161-8860-13e399d395f1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:25:05 crc kubenswrapper[4904]: E0214 11:25:05.491409 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-g77fw" podUID="1fc7a6f3-f356-4161-8860-13e399d395f1" Feb 14 11:25:05 crc kubenswrapper[4904]: E0214 11:25:05.502334 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 14 11:25:05 crc kubenswrapper[4904]: E0214 11:25:05.502719 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qhp84,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-gwbsl_openstack(28504613-d82c-4903-a641-684ab0e0c4c3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:25:05 crc kubenswrapper[4904]: E0214 11:25:05.503946 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" podUID="28504613-d82c-4903-a641-684ab0e0c4c3" Feb 14 11:25:05 crc kubenswrapper[4904]: E0214 11:25:05.509685 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 14 11:25:05 crc kubenswrapper[4904]: E0214 11:25:05.509885 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wp9f7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-kszpx_openstack(0688c0d9-6ac3-4c0e-a3d7-7745cd618c99): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:25:05 crc kubenswrapper[4904]: E0214 11:25:05.511194 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-kszpx" podUID="0688c0d9-6ac3-4c0e-a3d7-7745cd618c99" Feb 14 11:25:06 crc kubenswrapper[4904]: E0214 11:25:06.287513 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" podUID="0e9b3f7a-913c-43cd-bf74-a5b4df81567c" Feb 14 11:25:06 crc kubenswrapper[4904]: E0214 11:25:06.287718 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" podUID="28504613-d82c-4903-a641-684ab0e0c4c3" Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.329088 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-kszpx" event={"ID":"0688c0d9-6ac3-4c0e-a3d7-7745cd618c99","Type":"ContainerDied","Data":"a2f39eca9c6beaba2eb3b1a07b8d9f1185a54f4ef3fd290a6d07b31ce8ea7df3"} Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.329652 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2f39eca9c6beaba2eb3b1a07b8d9f1185a54f4ef3fd290a6d07b31ce8ea7df3" Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.330554 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-g77fw" Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.342322 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-kszpx" Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.352953 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-g77fw" event={"ID":"1fc7a6f3-f356-4161-8860-13e399d395f1","Type":"ContainerDied","Data":"fe0f60de7db8ec2afda7633117db4990d5f80f2c3cc1695b9f30ae68168c191b"} Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.479083 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0688c0d9-6ac3-4c0e-a3d7-7745cd618c99-dns-svc\") pod \"0688c0d9-6ac3-4c0e-a3d7-7745cd618c99\" (UID: \"0688c0d9-6ac3-4c0e-a3d7-7745cd618c99\") " Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.479264 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0688c0d9-6ac3-4c0e-a3d7-7745cd618c99-config\") pod \"0688c0d9-6ac3-4c0e-a3d7-7745cd618c99\" (UID: \"0688c0d9-6ac3-4c0e-a3d7-7745cd618c99\") " Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.479311 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp9f7\" (UniqueName: \"kubernetes.io/projected/0688c0d9-6ac3-4c0e-a3d7-7745cd618c99-kube-api-access-wp9f7\") pod \"0688c0d9-6ac3-4c0e-a3d7-7745cd618c99\" (UID: \"0688c0d9-6ac3-4c0e-a3d7-7745cd618c99\") " Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.479346 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fc7a6f3-f356-4161-8860-13e399d395f1-config\") pod \"1fc7a6f3-f356-4161-8860-13e399d395f1\" (UID: \"1fc7a6f3-f356-4161-8860-13e399d395f1\") " Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.479385 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6szmk\" (UniqueName: \"kubernetes.io/projected/1fc7a6f3-f356-4161-8860-13e399d395f1-kube-api-access-6szmk\") pod \"1fc7a6f3-f356-4161-8860-13e399d395f1\" (UID: \"1fc7a6f3-f356-4161-8860-13e399d395f1\") " Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.479999 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0688c0d9-6ac3-4c0e-a3d7-7745cd618c99-config" (OuterVolumeSpecName: "config") pod "0688c0d9-6ac3-4c0e-a3d7-7745cd618c99" (UID: "0688c0d9-6ac3-4c0e-a3d7-7745cd618c99"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.480364 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0688c0d9-6ac3-4c0e-a3d7-7745cd618c99-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0688c0d9-6ac3-4c0e-a3d7-7745cd618c99" (UID: "0688c0d9-6ac3-4c0e-a3d7-7745cd618c99"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.480880 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fc7a6f3-f356-4161-8860-13e399d395f1-config" (OuterVolumeSpecName: "config") pod "1fc7a6f3-f356-4161-8860-13e399d395f1" (UID: "1fc7a6f3-f356-4161-8860-13e399d395f1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.485953 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0688c0d9-6ac3-4c0e-a3d7-7745cd618c99-kube-api-access-wp9f7" (OuterVolumeSpecName: "kube-api-access-wp9f7") pod "0688c0d9-6ac3-4c0e-a3d7-7745cd618c99" (UID: "0688c0d9-6ac3-4c0e-a3d7-7745cd618c99"). InnerVolumeSpecName "kube-api-access-wp9f7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.487962 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fc7a6f3-f356-4161-8860-13e399d395f1-kube-api-access-6szmk" (OuterVolumeSpecName: "kube-api-access-6szmk") pod "1fc7a6f3-f356-4161-8860-13e399d395f1" (UID: "1fc7a6f3-f356-4161-8860-13e399d395f1"). InnerVolumeSpecName "kube-api-access-6szmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.577579 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.581468 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp9f7\" (UniqueName: \"kubernetes.io/projected/0688c0d9-6ac3-4c0e-a3d7-7745cd618c99-kube-api-access-wp9f7\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.581492 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fc7a6f3-f356-4161-8860-13e399d395f1-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.581506 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6szmk\" (UniqueName: \"kubernetes.io/projected/1fc7a6f3-f356-4161-8860-13e399d395f1-kube-api-access-6szmk\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.581516 4904 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0688c0d9-6ac3-4c0e-a3d7-7745cd618c99-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.581524 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0688c0d9-6ac3-4c0e-a3d7-7745cd618c99-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.584093 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 14 11:25:08 crc kubenswrapper[4904]: W0214 11:25:08.597087 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda35e40d9_4ced_4411_aff4_68c295331d47.slice/crio-35db55649b7c8fa081ade1c0094ef45e588f2bcd4d0ef143eb87b7d5931fef27 WatchSource:0}: Error finding container 35db55649b7c8fa081ade1c0094ef45e588f2bcd4d0ef143eb87b7d5931fef27: Status 404 returned error can't find the container with id 35db55649b7c8fa081ade1c0094ef45e588f2bcd4d0ef143eb87b7d5931fef27 Feb 14 11:25:08 crc kubenswrapper[4904]: W0214 11:25:08.598829 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6cb7af3_751c_48dd_84ae_6b77f7b9c233.slice/crio-50190847dd3fea122a70604f1c9ab433f05426aaeea219f4d7acf6083ce0625c WatchSource:0}: Error finding container 50190847dd3fea122a70604f1c9ab433f05426aaeea219f4d7acf6083ce0625c: Status 404 returned error can't find the container with id 50190847dd3fea122a70604f1c9ab433f05426aaeea219f4d7acf6083ce0625c Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.809908 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 14 11:25:08 crc kubenswrapper[4904]: I0214 11:25:08.817855 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zj52b"] Feb 14 11:25:08 crc kubenswrapper[4904]: W0214 11:25:08.966282 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05a401e2_caea_48a9_bb53_6a42b59dad57.slice/crio-486cde5bde2df33d93a24ffd4c431cdfce223fecdf07b3b179c59fcfedaea16f WatchSource:0}: Error finding container 486cde5bde2df33d93a24ffd4c431cdfce223fecdf07b3b179c59fcfedaea16f: Status 404 returned error can't find the container with id 486cde5bde2df33d93a24ffd4c431cdfce223fecdf07b3b179c59fcfedaea16f Feb 14 11:25:08 crc kubenswrapper[4904]: W0214 11:25:08.976417 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2878c0a_2515_4412_9aa7_acdb199bc463.slice/crio-5491d702b63eacc1b4a389a411f2ccde35375e0c776d6d3347ccefc37c9101e9 WatchSource:0}: Error finding container 5491d702b63eacc1b4a389a411f2ccde35375e0c776d6d3347ccefc37c9101e9: Status 404 returned error can't find the container with id 5491d702b63eacc1b4a389a411f2ccde35375e0c776d6d3347ccefc37c9101e9 Feb 14 11:25:09 crc kubenswrapper[4904]: I0214 11:25:09.101349 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 14 11:25:09 crc kubenswrapper[4904]: W0214 11:25:09.147161 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02df4ffa_81a6_4fd0_9136_3eedd479f653.slice/crio-d319fc8ee766a94d1e6bafc1f1a0c063955c3d3ca7573b75df229ad32cbeee76 WatchSource:0}: Error finding container d319fc8ee766a94d1e6bafc1f1a0c063955c3d3ca7573b75df229ad32cbeee76: Status 404 returned error can't find the container with id d319fc8ee766a94d1e6bafc1f1a0c063955c3d3ca7573b75df229ad32cbeee76 Feb 14 11:25:09 crc kubenswrapper[4904]: I0214 11:25:09.363303 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a35e40d9-4ced-4411-aff4-68c295331d47","Type":"ContainerStarted","Data":"35db55649b7c8fa081ade1c0094ef45e588f2bcd4d0ef143eb87b7d5931fef27"} Feb 14 11:25:09 crc kubenswrapper[4904]: I0214 11:25:09.365058 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"02df4ffa-81a6-4fd0-9136-3eedd479f653","Type":"ContainerStarted","Data":"d319fc8ee766a94d1e6bafc1f1a0c063955c3d3ca7573b75df229ad32cbeee76"} Feb 14 11:25:09 crc kubenswrapper[4904]: I0214 11:25:09.366779 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b6cb7af3-751c-48dd-84ae-6b77f7b9c233","Type":"ContainerStarted","Data":"50190847dd3fea122a70604f1c9ab433f05426aaeea219f4d7acf6083ce0625c"} Feb 14 11:25:09 crc kubenswrapper[4904]: I0214 11:25:09.367990 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zj52b" event={"ID":"05a401e2-caea-48a9-bb53-6a42b59dad57","Type":"ContainerStarted","Data":"486cde5bde2df33d93a24ffd4c431cdfce223fecdf07b3b179c59fcfedaea16f"} Feb 14 11:25:09 crc kubenswrapper[4904]: I0214 11:25:09.369415 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-kszpx" Feb 14 11:25:09 crc kubenswrapper[4904]: I0214 11:25:09.373176 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-g77fw" Feb 14 11:25:09 crc kubenswrapper[4904]: I0214 11:25:09.374506 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c2878c0a-2515-4412-9aa7-acdb199bc463","Type":"ContainerStarted","Data":"5491d702b63eacc1b4a389a411f2ccde35375e0c776d6d3347ccefc37c9101e9"} Feb 14 11:25:09 crc kubenswrapper[4904]: I0214 11:25:09.457246 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-g77fw"] Feb 14 11:25:09 crc kubenswrapper[4904]: I0214 11:25:09.464920 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-g77fw"] Feb 14 11:25:09 crc kubenswrapper[4904]: I0214 11:25:09.481328 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-kszpx"] Feb 14 11:25:09 crc kubenswrapper[4904]: I0214 11:25:09.489764 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-kszpx"] Feb 14 11:25:09 crc kubenswrapper[4904]: I0214 11:25:09.638927 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 14 11:25:09 crc kubenswrapper[4904]: I0214 11:25:09.856490 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0688c0d9-6ac3-4c0e-a3d7-7745cd618c99" path="/var/lib/kubelet/pods/0688c0d9-6ac3-4c0e-a3d7-7745cd618c99/volumes" Feb 14 11:25:09 crc kubenswrapper[4904]: I0214 11:25:09.860644 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fc7a6f3-f356-4161-8860-13e399d395f1" path="/var/lib/kubelet/pods/1fc7a6f3-f356-4161-8860-13e399d395f1/volumes" Feb 14 11:25:10 crc kubenswrapper[4904]: I0214 11:25:10.200862 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-9jk9m"] Feb 14 11:25:10 crc kubenswrapper[4904]: W0214 11:25:10.255889 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb07cea3_a6b2_45a1_8aba_fc64dbfa5c5c.slice/crio-b9d3e7a43bf321389d983b06848da3fc737b43834d5fd3573a5c45b2d1b164ea WatchSource:0}: Error finding container b9d3e7a43bf321389d983b06848da3fc737b43834d5fd3573a5c45b2d1b164ea: Status 404 returned error can't find the container with id b9d3e7a43bf321389d983b06848da3fc737b43834d5fd3573a5c45b2d1b164ea Feb 14 11:25:10 crc kubenswrapper[4904]: W0214 11:25:10.276207 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0422124f_385d_4e36_9a12_29bd9a6fa72e.slice/crio-32e08dac59a8630641424ec145f5b915034d1ab686b5333ac22d0ca6ab82b0c9 WatchSource:0}: Error finding container 32e08dac59a8630641424ec145f5b915034d1ab686b5333ac22d0ca6ab82b0c9: Status 404 returned error can't find the container with id 32e08dac59a8630641424ec145f5b915034d1ab686b5333ac22d0ca6ab82b0c9 Feb 14 11:25:10 crc kubenswrapper[4904]: I0214 11:25:10.385311 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9jk9m" event={"ID":"0422124f-385d-4e36-9a12-29bd9a6fa72e","Type":"ContainerStarted","Data":"32e08dac59a8630641424ec145f5b915034d1ab686b5333ac22d0ca6ab82b0c9"} Feb 14 11:25:10 crc kubenswrapper[4904]: I0214 11:25:10.386871 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7448d80b-d340-46a2-88a6-221db1f93f11","Type":"ContainerStarted","Data":"4e2f01e425ebbd9b9326491b905f29fd60a6633fb15da5c6854ae3eebc906136"} Feb 14 11:25:10 crc kubenswrapper[4904]: I0214 11:25:10.387959 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c","Type":"ContainerStarted","Data":"b9d3e7a43bf321389d983b06848da3fc737b43834d5fd3573a5c45b2d1b164ea"} Feb 14 11:25:13 crc kubenswrapper[4904]: I0214 11:25:13.407887 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2b06dd61-da9f-455f-b813-53300e4fbc10","Type":"ContainerStarted","Data":"67f4cf2e5ed352d8d50d8a7b06dbd4929f867edd15e88f56f90d69d36544af7b"} Feb 14 11:25:14 crc kubenswrapper[4904]: I0214 11:25:14.419811 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8515da2f-15c7-43af-b57f-bec14a6b0398","Type":"ContainerStarted","Data":"cabf8e56e34a6df0b7dde1428f2b80975722c3e1dc55210f459dd51cd7ab17a7"} Feb 14 11:25:14 crc kubenswrapper[4904]: I0214 11:25:14.420131 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 14 11:25:14 crc kubenswrapper[4904]: I0214 11:25:14.449538 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=16.816368823 podStartE2EDuration="23.44952296s" podCreationTimestamp="2026-02-14 11:24:51 +0000 UTC" firstStartedPulling="2026-02-14 11:25:04.628856919 +0000 UTC m=+895.441621600" lastFinishedPulling="2026-02-14 11:25:11.262011076 +0000 UTC m=+902.074775737" observedRunningTime="2026-02-14 11:25:14.438335261 +0000 UTC m=+905.251099922" watchObservedRunningTime="2026-02-14 11:25:14.44952296 +0000 UTC m=+905.262287621" Feb 14 11:25:15 crc kubenswrapper[4904]: I0214 11:25:15.826915 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wcd5n"] Feb 14 11:25:15 crc kubenswrapper[4904]: I0214 11:25:15.828666 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wcd5n" Feb 14 11:25:15 crc kubenswrapper[4904]: I0214 11:25:15.851121 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wcd5n"] Feb 14 11:25:15 crc kubenswrapper[4904]: I0214 11:25:15.911150 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmxgv\" (UniqueName: \"kubernetes.io/projected/28264935-4532-4d4e-b795-a8b927395d2e-kube-api-access-qmxgv\") pod \"community-operators-wcd5n\" (UID: \"28264935-4532-4d4e-b795-a8b927395d2e\") " pod="openshift-marketplace/community-operators-wcd5n" Feb 14 11:25:15 crc kubenswrapper[4904]: I0214 11:25:15.911290 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28264935-4532-4d4e-b795-a8b927395d2e-catalog-content\") pod \"community-operators-wcd5n\" (UID: \"28264935-4532-4d4e-b795-a8b927395d2e\") " pod="openshift-marketplace/community-operators-wcd5n" Feb 14 11:25:15 crc kubenswrapper[4904]: I0214 11:25:15.911330 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28264935-4532-4d4e-b795-a8b927395d2e-utilities\") pod \"community-operators-wcd5n\" (UID: \"28264935-4532-4d4e-b795-a8b927395d2e\") " pod="openshift-marketplace/community-operators-wcd5n" Feb 14 11:25:16 crc kubenswrapper[4904]: I0214 11:25:16.012548 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28264935-4532-4d4e-b795-a8b927395d2e-utilities\") pod \"community-operators-wcd5n\" (UID: \"28264935-4532-4d4e-b795-a8b927395d2e\") " pod="openshift-marketplace/community-operators-wcd5n" Feb 14 11:25:16 crc kubenswrapper[4904]: I0214 11:25:16.012643 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmxgv\" (UniqueName: \"kubernetes.io/projected/28264935-4532-4d4e-b795-a8b927395d2e-kube-api-access-qmxgv\") pod \"community-operators-wcd5n\" (UID: \"28264935-4532-4d4e-b795-a8b927395d2e\") " pod="openshift-marketplace/community-operators-wcd5n" Feb 14 11:25:16 crc kubenswrapper[4904]: I0214 11:25:16.012735 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28264935-4532-4d4e-b795-a8b927395d2e-catalog-content\") pod \"community-operators-wcd5n\" (UID: \"28264935-4532-4d4e-b795-a8b927395d2e\") " pod="openshift-marketplace/community-operators-wcd5n" Feb 14 11:25:16 crc kubenswrapper[4904]: I0214 11:25:16.013132 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28264935-4532-4d4e-b795-a8b927395d2e-utilities\") pod \"community-operators-wcd5n\" (UID: \"28264935-4532-4d4e-b795-a8b927395d2e\") " pod="openshift-marketplace/community-operators-wcd5n" Feb 14 11:25:16 crc kubenswrapper[4904]: I0214 11:25:16.013165 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28264935-4532-4d4e-b795-a8b927395d2e-catalog-content\") pod \"community-operators-wcd5n\" (UID: \"28264935-4532-4d4e-b795-a8b927395d2e\") " pod="openshift-marketplace/community-operators-wcd5n" Feb 14 11:25:16 crc kubenswrapper[4904]: I0214 11:25:16.032291 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmxgv\" (UniqueName: \"kubernetes.io/projected/28264935-4532-4d4e-b795-a8b927395d2e-kube-api-access-qmxgv\") pod \"community-operators-wcd5n\" (UID: \"28264935-4532-4d4e-b795-a8b927395d2e\") " pod="openshift-marketplace/community-operators-wcd5n" Feb 14 11:25:16 crc kubenswrapper[4904]: I0214 11:25:16.155666 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wcd5n" Feb 14 11:25:16 crc kubenswrapper[4904]: I0214 11:25:16.382717 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:25:16 crc kubenswrapper[4904]: I0214 11:25:16.382785 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:25:21 crc kubenswrapper[4904]: I0214 11:25:21.720554 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wcd5n"] Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.292023 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.489687 4904 generic.go:334] "Generic (PLEG): container finished" podID="28264935-4532-4d4e-b795-a8b927395d2e" containerID="139aa32e3123828d9af3e7d70b1422d3372f7eef680e428d808c6f4780ae6c75" exitCode=0 Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.489752 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wcd5n" event={"ID":"28264935-4532-4d4e-b795-a8b927395d2e","Type":"ContainerDied","Data":"139aa32e3123828d9af3e7d70b1422d3372f7eef680e428d808c6f4780ae6c75"} Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.489777 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wcd5n" event={"ID":"28264935-4532-4d4e-b795-a8b927395d2e","Type":"ContainerStarted","Data":"d72b16ce3deb53adfb9e7dbdb8e5e8fd7b5e77858a17f4950cd3a121e992e618"} Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.493668 4904 generic.go:334] "Generic (PLEG): container finished" podID="0422124f-385d-4e36-9a12-29bd9a6fa72e" containerID="d5428132c5d33abea7c91fda4c44ace33165d80ccffae3c25b3282ad5e9f6698" exitCode=0 Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.493708 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9jk9m" event={"ID":"0422124f-385d-4e36-9a12-29bd9a6fa72e","Type":"ContainerDied","Data":"d5428132c5d33abea7c91fda4c44ace33165d80ccffae3c25b3282ad5e9f6698"} Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.499745 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b6cb7af3-751c-48dd-84ae-6b77f7b9c233","Type":"ContainerStarted","Data":"f900c85d63a53d8495e9d2696b79a3be737fbd74527870c0a3846da9768f42a2"} Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.501248 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c","Type":"ContainerStarted","Data":"e8f3f0b56e5ed1831d4febf3eadb7459a53383a899cd8ead7e2c51427023700c"} Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.508810 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zj52b" event={"ID":"05a401e2-caea-48a9-bb53-6a42b59dad57","Type":"ContainerStarted","Data":"b0bf9228faac2178fe32c9cb97cc352b2f386e7367430b23e8b2bc461f5ce943"} Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.509670 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-zj52b" Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.512394 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c2878c0a-2515-4412-9aa7-acdb199bc463","Type":"ContainerStarted","Data":"f09a01ec62183f69093346403ed195c30f71b7ee36294bd15a4df1023d3df47a"} Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.514772 4904 generic.go:334] "Generic (PLEG): container finished" podID="0e9b3f7a-913c-43cd-bf74-a5b4df81567c" containerID="64b1382e277f66b1de49cccc115892dd00d8679ffa470cb5ec953346a505cf0c" exitCode=0 Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.514803 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" event={"ID":"0e9b3f7a-913c-43cd-bf74-a5b4df81567c","Type":"ContainerDied","Data":"64b1382e277f66b1de49cccc115892dd00d8679ffa470cb5ec953346a505cf0c"} Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.516751 4904 generic.go:334] "Generic (PLEG): container finished" podID="28504613-d82c-4903-a641-684ab0e0c4c3" containerID="3e5412c8e0e1765fa07fc77a173d41d7eea87fbbf719ea772ce273474a24a58f" exitCode=0 Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.516813 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" event={"ID":"28504613-d82c-4903-a641-684ab0e0c4c3","Type":"ContainerDied","Data":"3e5412c8e0e1765fa07fc77a173d41d7eea87fbbf719ea772ce273474a24a58f"} Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.519572 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a35e40d9-4ced-4411-aff4-68c295331d47","Type":"ContainerStarted","Data":"94db504293d536f52a589697f381202b5c653280df4e6cbbe7e5a53ace97146a"} Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.519646 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.521036 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"02df4ffa-81a6-4fd0-9136-3eedd479f653","Type":"ContainerStarted","Data":"cee6971f3635fb0276a9e335bf66e00a4c2a3fc1c96f0bd154da1a0f7f9bc8ce"} Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.587437 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=30.926188248 podStartE2EDuration="33.587389236s" podCreationTimestamp="2026-02-14 11:24:49 +0000 UTC" firstStartedPulling="2026-02-14 11:25:08.600820798 +0000 UTC m=+899.413585459" lastFinishedPulling="2026-02-14 11:25:11.262021786 +0000 UTC m=+902.074786447" observedRunningTime="2026-02-14 11:25:22.567238634 +0000 UTC m=+913.380003305" watchObservedRunningTime="2026-02-14 11:25:22.587389236 +0000 UTC m=+913.400153897" Feb 14 11:25:22 crc kubenswrapper[4904]: I0214 11:25:22.647356 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-zj52b" podStartSLOduration=16.256980036 podStartE2EDuration="28.64733719s" podCreationTimestamp="2026-02-14 11:24:54 +0000 UTC" firstStartedPulling="2026-02-14 11:25:08.970657097 +0000 UTC m=+899.783421758" lastFinishedPulling="2026-02-14 11:25:21.361014251 +0000 UTC m=+912.173778912" observedRunningTime="2026-02-14 11:25:22.641962343 +0000 UTC m=+913.454727004" watchObservedRunningTime="2026-02-14 11:25:22.64733719 +0000 UTC m=+913.460101841" Feb 14 11:25:23 crc kubenswrapper[4904]: I0214 11:25:23.528782 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9jk9m" event={"ID":"0422124f-385d-4e36-9a12-29bd9a6fa72e","Type":"ContainerStarted","Data":"92dc4d03bf4d4f56b8bd15aa3020fcdf30c394023acd36c34a1c12d71191faef"} Feb 14 11:25:23 crc kubenswrapper[4904]: I0214 11:25:23.536100 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" event={"ID":"28504613-d82c-4903-a641-684ab0e0c4c3","Type":"ContainerStarted","Data":"e46ea0a2ca964241bba0d7217570388fd11c6168f7bf7b443d3e848a341399c3"} Feb 14 11:25:23 crc kubenswrapper[4904]: I0214 11:25:23.536258 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" Feb 14 11:25:23 crc kubenswrapper[4904]: I0214 11:25:23.538104 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"02df4ffa-81a6-4fd0-9136-3eedd479f653","Type":"ContainerStarted","Data":"b3527fdd177818e0bac22b568c270b4cdab1fcea93a710a580ebec22c598adff"} Feb 14 11:25:23 crc kubenswrapper[4904]: I0214 11:25:23.540911 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c","Type":"ContainerStarted","Data":"fc0bc1b1b97beb77fbb8201a6efcec871f2c7ddd7708436364b57ea99ce5df7c"} Feb 14 11:25:23 crc kubenswrapper[4904]: I0214 11:25:23.543025 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" event={"ID":"0e9b3f7a-913c-43cd-bf74-a5b4df81567c","Type":"ContainerStarted","Data":"2cc7c942f5624ea094277e37a3898d5df3f8e3712d171533a7c2d4b5e060dee9"} Feb 14 11:25:23 crc kubenswrapper[4904]: I0214 11:25:23.556641 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" podStartSLOduration=3.483245823 podStartE2EDuration="38.556625562s" podCreationTimestamp="2026-02-14 11:24:45 +0000 UTC" firstStartedPulling="2026-02-14 11:24:46.64350296 +0000 UTC m=+877.456267621" lastFinishedPulling="2026-02-14 11:25:21.716882709 +0000 UTC m=+912.529647360" observedRunningTime="2026-02-14 11:25:23.553904087 +0000 UTC m=+914.366668768" watchObservedRunningTime="2026-02-14 11:25:23.556625562 +0000 UTC m=+914.369390223" Feb 14 11:25:23 crc kubenswrapper[4904]: I0214 11:25:23.629161 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=15.740044987 podStartE2EDuration="29.62914345s" podCreationTimestamp="2026-02-14 11:24:54 +0000 UTC" firstStartedPulling="2026-02-14 11:25:09.154674987 +0000 UTC m=+899.967439648" lastFinishedPulling="2026-02-14 11:25:23.04377345 +0000 UTC m=+913.856538111" observedRunningTime="2026-02-14 11:25:23.611286841 +0000 UTC m=+914.424051502" watchObservedRunningTime="2026-02-14 11:25:23.62914345 +0000 UTC m=+914.441908101" Feb 14 11:25:23 crc kubenswrapper[4904]: I0214 11:25:23.637464 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=13.853178361 podStartE2EDuration="26.637448287s" podCreationTimestamp="2026-02-14 11:24:57 +0000 UTC" firstStartedPulling="2026-02-14 11:25:10.259535094 +0000 UTC m=+901.072299755" lastFinishedPulling="2026-02-14 11:25:23.04380502 +0000 UTC m=+913.856569681" observedRunningTime="2026-02-14 11:25:23.626514318 +0000 UTC m=+914.439278979" watchObservedRunningTime="2026-02-14 11:25:23.637448287 +0000 UTC m=+914.450212948" Feb 14 11:25:23 crc kubenswrapper[4904]: I0214 11:25:23.645469 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" podStartSLOduration=3.480316585 podStartE2EDuration="38.645447597s" podCreationTimestamp="2026-02-14 11:24:45 +0000 UTC" firstStartedPulling="2026-02-14 11:24:46.553243518 +0000 UTC m=+877.366008179" lastFinishedPulling="2026-02-14 11:25:21.71837453 +0000 UTC m=+912.531139191" observedRunningTime="2026-02-14 11:25:23.640732818 +0000 UTC m=+914.453497479" watchObservedRunningTime="2026-02-14 11:25:23.645447597 +0000 UTC m=+914.458212258" Feb 14 11:25:23 crc kubenswrapper[4904]: I0214 11:25:23.826152 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Feb 14 11:25:24 crc kubenswrapper[4904]: I0214 11:25:24.552081 4904 generic.go:334] "Generic (PLEG): container finished" podID="28264935-4532-4d4e-b795-a8b927395d2e" containerID="eb833188011ca621a3318aa7c5b10a299e65eb1586a7f5f5fe9af7d6f09cc32c" exitCode=0 Feb 14 11:25:24 crc kubenswrapper[4904]: I0214 11:25:24.552210 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wcd5n" event={"ID":"28264935-4532-4d4e-b795-a8b927395d2e","Type":"ContainerDied","Data":"eb833188011ca621a3318aa7c5b10a299e65eb1586a7f5f5fe9af7d6f09cc32c"} Feb 14 11:25:24 crc kubenswrapper[4904]: I0214 11:25:24.555295 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9jk9m" event={"ID":"0422124f-385d-4e36-9a12-29bd9a6fa72e","Type":"ContainerStarted","Data":"8e9f337fb740e9d5cb4d4cf3553fd5d3af5dc181579d7ac2ed47307d7ecbb0be"} Feb 14 11:25:24 crc kubenswrapper[4904]: I0214 11:25:24.556128 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:25:24 crc kubenswrapper[4904]: I0214 11:25:24.598302 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-9jk9m" podStartSLOduration=19.619173803 podStartE2EDuration="30.598280072s" podCreationTimestamp="2026-02-14 11:24:54 +0000 UTC" firstStartedPulling="2026-02-14 11:25:10.28113881 +0000 UTC m=+901.093903471" lastFinishedPulling="2026-02-14 11:25:21.260245079 +0000 UTC m=+912.073009740" observedRunningTime="2026-02-14 11:25:24.594422996 +0000 UTC m=+915.407187647" watchObservedRunningTime="2026-02-14 11:25:24.598280072 +0000 UTC m=+915.411044733" Feb 14 11:25:24 crc kubenswrapper[4904]: I0214 11:25:24.964682 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:25:25 crc kubenswrapper[4904]: I0214 11:25:25.563986 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wcd5n" event={"ID":"28264935-4532-4d4e-b795-a8b927395d2e","Type":"ContainerStarted","Data":"067ab3afa3a539abf7f0361c03fe3e66c38b9aba0764cdaa8573cd07f2b5994f"} Feb 14 11:25:25 crc kubenswrapper[4904]: I0214 11:25:25.598182 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wcd5n" podStartSLOduration=8.514937148 podStartE2EDuration="10.598162977s" podCreationTimestamp="2026-02-14 11:25:15 +0000 UTC" firstStartedPulling="2026-02-14 11:25:22.925468676 +0000 UTC m=+913.738233337" lastFinishedPulling="2026-02-14 11:25:25.008694505 +0000 UTC m=+915.821459166" observedRunningTime="2026-02-14 11:25:25.596390227 +0000 UTC m=+916.409154888" watchObservedRunningTime="2026-02-14 11:25:25.598162977 +0000 UTC m=+916.410927638" Feb 14 11:25:25 crc kubenswrapper[4904]: I0214 11:25:25.825650 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Feb 14 11:25:25 crc kubenswrapper[4904]: I0214 11:25:25.855205 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" Feb 14 11:25:25 crc kubenswrapper[4904]: I0214 11:25:25.868436 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.156895 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wcd5n" Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.156990 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wcd5n" Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.314404 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.314446 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.356453 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.571668 4904 generic.go:334] "Generic (PLEG): container finished" podID="c2878c0a-2515-4412-9aa7-acdb199bc463" containerID="f09a01ec62183f69093346403ed195c30f71b7ee36294bd15a4df1023d3df47a" exitCode=0 Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.571737 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c2878c0a-2515-4412-9aa7-acdb199bc463","Type":"ContainerDied","Data":"f09a01ec62183f69093346403ed195c30f71b7ee36294bd15a4df1023d3df47a"} Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.574590 4904 generic.go:334] "Generic (PLEG): container finished" podID="b6cb7af3-751c-48dd-84ae-6b77f7b9c233" containerID="f900c85d63a53d8495e9d2696b79a3be737fbd74527870c0a3846da9768f42a2" exitCode=0 Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.574695 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b6cb7af3-751c-48dd-84ae-6b77f7b9c233","Type":"ContainerDied","Data":"f900c85d63a53d8495e9d2696b79a3be737fbd74527870c0a3846da9768f42a2"} Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.627612 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.690771 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.903320 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-gwbsl"] Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.903526 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" podUID="28504613-d82c-4903-a641-684ab0e0c4c3" containerName="dnsmasq-dns" containerID="cri-o://e46ea0a2ca964241bba0d7217570388fd11c6168f7bf7b443d3e848a341399c3" gracePeriod=10 Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.946637 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-5vxdt"] Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.947786 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.951255 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.968623 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-5vxdt"] Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.990676 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e75fcff-e55b-43cd-97fb-2c1eeb565168-config\") pod \"dnsmasq-dns-7f896c8c65-5vxdt\" (UID: \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\") " pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.990739 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq8pg\" (UniqueName: \"kubernetes.io/projected/3e75fcff-e55b-43cd-97fb-2c1eeb565168-kube-api-access-jq8pg\") pod \"dnsmasq-dns-7f896c8c65-5vxdt\" (UID: \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\") " pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.990813 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e75fcff-e55b-43cd-97fb-2c1eeb565168-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-5vxdt\" (UID: \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\") " pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" Feb 14 11:25:26 crc kubenswrapper[4904]: I0214 11:25:26.990876 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e75fcff-e55b-43cd-97fb-2c1eeb565168-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-5vxdt\" (UID: \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\") " pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.092331 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e75fcff-e55b-43cd-97fb-2c1eeb565168-config\") pod \"dnsmasq-dns-7f896c8c65-5vxdt\" (UID: \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\") " pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.092393 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq8pg\" (UniqueName: \"kubernetes.io/projected/3e75fcff-e55b-43cd-97fb-2c1eeb565168-kube-api-access-jq8pg\") pod \"dnsmasq-dns-7f896c8c65-5vxdt\" (UID: \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\") " pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.092435 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e75fcff-e55b-43cd-97fb-2c1eeb565168-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-5vxdt\" (UID: \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\") " pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.092457 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e75fcff-e55b-43cd-97fb-2c1eeb565168-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-5vxdt\" (UID: \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\") " pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.093540 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e75fcff-e55b-43cd-97fb-2c1eeb565168-config\") pod \"dnsmasq-dns-7f896c8c65-5vxdt\" (UID: \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\") " pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.093687 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e75fcff-e55b-43cd-97fb-2c1eeb565168-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-5vxdt\" (UID: \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\") " pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.093708 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e75fcff-e55b-43cd-97fb-2c1eeb565168-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-5vxdt\" (UID: \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\") " pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.112610 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq8pg\" (UniqueName: \"kubernetes.io/projected/3e75fcff-e55b-43cd-97fb-2c1eeb565168-kube-api-access-jq8pg\") pod \"dnsmasq-dns-7f896c8c65-5vxdt\" (UID: \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\") " pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.151976 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-s87wf"] Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.152973 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.159079 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.159110 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-s87wf"] Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.195422 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/692d42dd-2dae-4d7f-bd38-08720eb781bc-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-s87wf\" (UID: \"692d42dd-2dae-4d7f-bd38-08720eb781bc\") " pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.195461 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/692d42dd-2dae-4d7f-bd38-08720eb781bc-combined-ca-bundle\") pod \"ovn-controller-metrics-s87wf\" (UID: \"692d42dd-2dae-4d7f-bd38-08720eb781bc\") " pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.195529 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvnhd\" (UniqueName: \"kubernetes.io/projected/692d42dd-2dae-4d7f-bd38-08720eb781bc-kube-api-access-mvnhd\") pod \"ovn-controller-metrics-s87wf\" (UID: \"692d42dd-2dae-4d7f-bd38-08720eb781bc\") " pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.195564 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/692d42dd-2dae-4d7f-bd38-08720eb781bc-config\") pod \"ovn-controller-metrics-s87wf\" (UID: \"692d42dd-2dae-4d7f-bd38-08720eb781bc\") " pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.195585 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/692d42dd-2dae-4d7f-bd38-08720eb781bc-ovs-rundir\") pod \"ovn-controller-metrics-s87wf\" (UID: \"692d42dd-2dae-4d7f-bd38-08720eb781bc\") " pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.195614 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/692d42dd-2dae-4d7f-bd38-08720eb781bc-ovn-rundir\") pod \"ovn-controller-metrics-s87wf\" (UID: \"692d42dd-2dae-4d7f-bd38-08720eb781bc\") " pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.215495 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-wcd5n" podUID="28264935-4532-4d4e-b795-a8b927395d2e" containerName="registry-server" probeResult="failure" output=< Feb 14 11:25:27 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 11:25:27 crc kubenswrapper[4904]: > Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.247014 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.248248 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.266182 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.268743 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-659zv" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.269136 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.269569 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.269813 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.292226 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.302619 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/eed3ea68-257f-49bd-8a2b-4f86a34590c9-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.302666 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eed3ea68-257f-49bd-8a2b-4f86a34590c9-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.302696 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/692d42dd-2dae-4d7f-bd38-08720eb781bc-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-s87wf\" (UID: \"692d42dd-2dae-4d7f-bd38-08720eb781bc\") " pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.302716 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/692d42dd-2dae-4d7f-bd38-08720eb781bc-combined-ca-bundle\") pod \"ovn-controller-metrics-s87wf\" (UID: \"692d42dd-2dae-4d7f-bd38-08720eb781bc\") " pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.302764 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vs8n5\" (UniqueName: \"kubernetes.io/projected/eed3ea68-257f-49bd-8a2b-4f86a34590c9-kube-api-access-vs8n5\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.302804 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvnhd\" (UniqueName: \"kubernetes.io/projected/692d42dd-2dae-4d7f-bd38-08720eb781bc-kube-api-access-mvnhd\") pod \"ovn-controller-metrics-s87wf\" (UID: \"692d42dd-2dae-4d7f-bd38-08720eb781bc\") " pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.302841 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/692d42dd-2dae-4d7f-bd38-08720eb781bc-config\") pod \"ovn-controller-metrics-s87wf\" (UID: \"692d42dd-2dae-4d7f-bd38-08720eb781bc\") " pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.302858 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eed3ea68-257f-49bd-8a2b-4f86a34590c9-scripts\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.302878 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eed3ea68-257f-49bd-8a2b-4f86a34590c9-config\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.302892 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/eed3ea68-257f-49bd-8a2b-4f86a34590c9-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.302908 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/692d42dd-2dae-4d7f-bd38-08720eb781bc-ovs-rundir\") pod \"ovn-controller-metrics-s87wf\" (UID: \"692d42dd-2dae-4d7f-bd38-08720eb781bc\") " pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.302931 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/eed3ea68-257f-49bd-8a2b-4f86a34590c9-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.302949 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/692d42dd-2dae-4d7f-bd38-08720eb781bc-ovn-rundir\") pod \"ovn-controller-metrics-s87wf\" (UID: \"692d42dd-2dae-4d7f-bd38-08720eb781bc\") " pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.303158 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/692d42dd-2dae-4d7f-bd38-08720eb781bc-ovn-rundir\") pod \"ovn-controller-metrics-s87wf\" (UID: \"692d42dd-2dae-4d7f-bd38-08720eb781bc\") " pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.307062 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/692d42dd-2dae-4d7f-bd38-08720eb781bc-config\") pod \"ovn-controller-metrics-s87wf\" (UID: \"692d42dd-2dae-4d7f-bd38-08720eb781bc\") " pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.307145 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/692d42dd-2dae-4d7f-bd38-08720eb781bc-ovs-rundir\") pod \"ovn-controller-metrics-s87wf\" (UID: \"692d42dd-2dae-4d7f-bd38-08720eb781bc\") " pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.312193 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/692d42dd-2dae-4d7f-bd38-08720eb781bc-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-s87wf\" (UID: \"692d42dd-2dae-4d7f-bd38-08720eb781bc\") " pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.315669 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/692d42dd-2dae-4d7f-bd38-08720eb781bc-combined-ca-bundle\") pod \"ovn-controller-metrics-s87wf\" (UID: \"692d42dd-2dae-4d7f-bd38-08720eb781bc\") " pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.324624 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qhdgb"] Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.325192 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" podUID="0e9b3f7a-913c-43cd-bf74-a5b4df81567c" containerName="dnsmasq-dns" containerID="cri-o://2cc7c942f5624ea094277e37a3898d5df3f8e3712d171533a7c2d4b5e060dee9" gracePeriod=10 Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.344612 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvnhd\" (UniqueName: \"kubernetes.io/projected/692d42dd-2dae-4d7f-bd38-08720eb781bc-kube-api-access-mvnhd\") pod \"ovn-controller-metrics-s87wf\" (UID: \"692d42dd-2dae-4d7f-bd38-08720eb781bc\") " pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.380903 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-2db8z"] Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.382402 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.386676 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.406539 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/eed3ea68-257f-49bd-8a2b-4f86a34590c9-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.406688 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/eed3ea68-257f-49bd-8a2b-4f86a34590c9-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.406769 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eed3ea68-257f-49bd-8a2b-4f86a34590c9-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.406900 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vs8n5\" (UniqueName: \"kubernetes.io/projected/eed3ea68-257f-49bd-8a2b-4f86a34590c9-kube-api-access-vs8n5\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.407029 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eed3ea68-257f-49bd-8a2b-4f86a34590c9-scripts\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.407101 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eed3ea68-257f-49bd-8a2b-4f86a34590c9-config\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.407184 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/eed3ea68-257f-49bd-8a2b-4f86a34590c9-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.410175 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eed3ea68-257f-49bd-8a2b-4f86a34590c9-scripts\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.411022 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eed3ea68-257f-49bd-8a2b-4f86a34590c9-config\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.411268 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/eed3ea68-257f-49bd-8a2b-4f86a34590c9-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.411300 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-2db8z"] Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.413223 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/eed3ea68-257f-49bd-8a2b-4f86a34590c9-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.415479 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/eed3ea68-257f-49bd-8a2b-4f86a34590c9-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.416526 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eed3ea68-257f-49bd-8a2b-4f86a34590c9-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.439355 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vs8n5\" (UniqueName: \"kubernetes.io/projected/eed3ea68-257f-49bd-8a2b-4f86a34590c9-kube-api-access-vs8n5\") pod \"ovn-northd-0\" (UID: \"eed3ea68-257f-49bd-8a2b-4f86a34590c9\") " pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.474445 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-s87wf" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.508202 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-config\") pod \"dnsmasq-dns-86db49b7ff-2db8z\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.508267 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-2db8z\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.508287 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-2db8z\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.508313 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvrpk\" (UniqueName: \"kubernetes.io/projected/b6914d96-01d9-45d6-933a-b3b8875349df-kube-api-access-mvrpk\") pod \"dnsmasq-dns-86db49b7ff-2db8z\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.508388 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-2db8z\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.521204 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.616245 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.618505 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28504613-d82c-4903-a641-684ab0e0c4c3-dns-svc\") pod \"28504613-d82c-4903-a641-684ab0e0c4c3\" (UID: \"28504613-d82c-4903-a641-684ab0e0c4c3\") " Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.618739 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhp84\" (UniqueName: \"kubernetes.io/projected/28504613-d82c-4903-a641-684ab0e0c4c3-kube-api-access-qhp84\") pod \"28504613-d82c-4903-a641-684ab0e0c4c3\" (UID: \"28504613-d82c-4903-a641-684ab0e0c4c3\") " Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.618913 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28504613-d82c-4903-a641-684ab0e0c4c3-config\") pod \"28504613-d82c-4903-a641-684ab0e0c4c3\" (UID: \"28504613-d82c-4903-a641-684ab0e0c4c3\") " Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.622790 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-config\") pod \"dnsmasq-dns-86db49b7ff-2db8z\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.622849 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-2db8z\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.622924 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-2db8z\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.622990 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvrpk\" (UniqueName: \"kubernetes.io/projected/b6914d96-01d9-45d6-933a-b3b8875349df-kube-api-access-mvrpk\") pod \"dnsmasq-dns-86db49b7ff-2db8z\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.623204 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-2db8z\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.624568 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-2db8z\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.627365 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-2db8z\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.631829 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28504613-d82c-4903-a641-684ab0e0c4c3-kube-api-access-qhp84" (OuterVolumeSpecName: "kube-api-access-qhp84") pod "28504613-d82c-4903-a641-684ab0e0c4c3" (UID: "28504613-d82c-4903-a641-684ab0e0c4c3"). InnerVolumeSpecName "kube-api-access-qhp84". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.632684 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-2db8z\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.634863 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-config\") pod \"dnsmasq-dns-86db49b7ff-2db8z\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.656079 4904 generic.go:334] "Generic (PLEG): container finished" podID="28504613-d82c-4903-a641-684ab0e0c4c3" containerID="e46ea0a2ca964241bba0d7217570388fd11c6168f7bf7b443d3e848a341399c3" exitCode=0 Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.656530 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.657510 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" event={"ID":"28504613-d82c-4903-a641-684ab0e0c4c3","Type":"ContainerDied","Data":"e46ea0a2ca964241bba0d7217570388fd11c6168f7bf7b443d3e848a341399c3"} Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.657668 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-gwbsl" event={"ID":"28504613-d82c-4903-a641-684ab0e0c4c3","Type":"ContainerDied","Data":"8e1380ab4cd03cfd5371eb576851f1c94c2940feca272a5eb6a42c293c3a2abc"} Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.652635 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvrpk\" (UniqueName: \"kubernetes.io/projected/b6914d96-01d9-45d6-933a-b3b8875349df-kube-api-access-mvrpk\") pod \"dnsmasq-dns-86db49b7ff-2db8z\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.658467 4904 scope.go:117] "RemoveContainer" containerID="e46ea0a2ca964241bba0d7217570388fd11c6168f7bf7b443d3e848a341399c3" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.728621 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhp84\" (UniqueName: \"kubernetes.io/projected/28504613-d82c-4903-a641-684ab0e0c4c3-kube-api-access-qhp84\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.729898 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b6cb7af3-751c-48dd-84ae-6b77f7b9c233","Type":"ContainerStarted","Data":"b90d51ef8677f26d2c4b7a9cf1a8c05f4c36327614fe80c844a09f8ba26a41c4"} Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.729997 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28504613-d82c-4903-a641-684ab0e0c4c3-config" (OuterVolumeSpecName: "config") pod "28504613-d82c-4903-a641-684ab0e0c4c3" (UID: "28504613-d82c-4903-a641-684ab0e0c4c3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.750795 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c2878c0a-2515-4412-9aa7-acdb199bc463","Type":"ContainerStarted","Data":"864e9d2be182a1b1e9a0737175911bc1cab7a01d12e92b2eddc0936906014f7c"} Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.776170 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.786179 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28504613-d82c-4903-a641-684ab0e0c4c3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "28504613-d82c-4903-a641-684ab0e0c4c3" (UID: "28504613-d82c-4903-a641-684ab0e0c4c3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.793257 4904 generic.go:334] "Generic (PLEG): container finished" podID="0e9b3f7a-913c-43cd-bf74-a5b4df81567c" containerID="2cc7c942f5624ea094277e37a3898d5df3f8e3712d171533a7c2d4b5e060dee9" exitCode=0 Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.793999 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=27.515379073 podStartE2EDuration="39.793983942s" podCreationTimestamp="2026-02-14 11:24:48 +0000 UTC" firstStartedPulling="2026-02-14 11:25:08.98163892 +0000 UTC m=+899.794403581" lastFinishedPulling="2026-02-14 11:25:21.260243789 +0000 UTC m=+912.073008450" observedRunningTime="2026-02-14 11:25:27.788283826 +0000 UTC m=+918.601048487" watchObservedRunningTime="2026-02-14 11:25:27.793983942 +0000 UTC m=+918.606748603" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.794182 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" event={"ID":"0e9b3f7a-913c-43cd-bf74-a5b4df81567c","Type":"ContainerDied","Data":"2cc7c942f5624ea094277e37a3898d5df3f8e3712d171533a7c2d4b5e060dee9"} Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.795090 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=28.169697636 podStartE2EDuration="40.795084972s" podCreationTimestamp="2026-02-14 11:24:47 +0000 UTC" firstStartedPulling="2026-02-14 11:25:08.604403428 +0000 UTC m=+899.417168089" lastFinishedPulling="2026-02-14 11:25:21.229790764 +0000 UTC m=+912.042555425" observedRunningTime="2026-02-14 11:25:27.754977752 +0000 UTC m=+918.567742413" watchObservedRunningTime="2026-02-14 11:25:27.795084972 +0000 UTC m=+918.607849633" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.835130 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28504613-d82c-4903-a641-684ab0e0c4c3-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.835231 4904 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28504613-d82c-4903-a641-684ab0e0c4c3-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.846039 4904 scope.go:117] "RemoveContainer" containerID="3e5412c8e0e1765fa07fc77a173d41d7eea87fbbf719ea772ce273474a24a58f" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.887652 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-5vxdt"] Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.921217 4904 scope.go:117] "RemoveContainer" containerID="e46ea0a2ca964241bba0d7217570388fd11c6168f7bf7b443d3e848a341399c3" Feb 14 11:25:27 crc kubenswrapper[4904]: E0214 11:25:27.923072 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e46ea0a2ca964241bba0d7217570388fd11c6168f7bf7b443d3e848a341399c3\": container with ID starting with e46ea0a2ca964241bba0d7217570388fd11c6168f7bf7b443d3e848a341399c3 not found: ID does not exist" containerID="e46ea0a2ca964241bba0d7217570388fd11c6168f7bf7b443d3e848a341399c3" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.923099 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e46ea0a2ca964241bba0d7217570388fd11c6168f7bf7b443d3e848a341399c3"} err="failed to get container status \"e46ea0a2ca964241bba0d7217570388fd11c6168f7bf7b443d3e848a341399c3\": rpc error: code = NotFound desc = could not find container \"e46ea0a2ca964241bba0d7217570388fd11c6168f7bf7b443d3e848a341399c3\": container with ID starting with e46ea0a2ca964241bba0d7217570388fd11c6168f7bf7b443d3e848a341399c3 not found: ID does not exist" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.923119 4904 scope.go:117] "RemoveContainer" containerID="3e5412c8e0e1765fa07fc77a173d41d7eea87fbbf719ea772ce273474a24a58f" Feb 14 11:25:27 crc kubenswrapper[4904]: E0214 11:25:27.929889 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e5412c8e0e1765fa07fc77a173d41d7eea87fbbf719ea772ce273474a24a58f\": container with ID starting with 3e5412c8e0e1765fa07fc77a173d41d7eea87fbbf719ea772ce273474a24a58f not found: ID does not exist" containerID="3e5412c8e0e1765fa07fc77a173d41d7eea87fbbf719ea772ce273474a24a58f" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.929920 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e5412c8e0e1765fa07fc77a173d41d7eea87fbbf719ea772ce273474a24a58f"} err="failed to get container status \"3e5412c8e0e1765fa07fc77a173d41d7eea87fbbf719ea772ce273474a24a58f\": rpc error: code = NotFound desc = could not find container \"3e5412c8e0e1765fa07fc77a173d41d7eea87fbbf719ea772ce273474a24a58f\": container with ID starting with 3e5412c8e0e1765fa07fc77a173d41d7eea87fbbf719ea772ce273474a24a58f not found: ID does not exist" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.991683 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-gwbsl"] Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.992303 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" Feb 14 11:25:27 crc kubenswrapper[4904]: I0214 11:25:27.998385 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-gwbsl"] Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.145368 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-config\") pod \"0e9b3f7a-913c-43cd-bf74-a5b4df81567c\" (UID: \"0e9b3f7a-913c-43cd-bf74-a5b4df81567c\") " Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.145816 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v99fl\" (UniqueName: \"kubernetes.io/projected/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-kube-api-access-v99fl\") pod \"0e9b3f7a-913c-43cd-bf74-a5b4df81567c\" (UID: \"0e9b3f7a-913c-43cd-bf74-a5b4df81567c\") " Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.146070 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-dns-svc\") pod \"0e9b3f7a-913c-43cd-bf74-a5b4df81567c\" (UID: \"0e9b3f7a-913c-43cd-bf74-a5b4df81567c\") " Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.151954 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-kube-api-access-v99fl" (OuterVolumeSpecName: "kube-api-access-v99fl") pod "0e9b3f7a-913c-43cd-bf74-a5b4df81567c" (UID: "0e9b3f7a-913c-43cd-bf74-a5b4df81567c"). InnerVolumeSpecName "kube-api-access-v99fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:25:28 crc kubenswrapper[4904]: E0214 11:25:28.190323 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-dns-svc podName:0e9b3f7a-913c-43cd-bf74-a5b4df81567c nodeName:}" failed. No retries permitted until 2026-02-14 11:25:28.690299958 +0000 UTC m=+919.503064619 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "dns-svc" (UniqueName: "kubernetes.io/configmap/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-dns-svc") pod "0e9b3f7a-913c-43cd-bf74-a5b4df81567c" (UID: "0e9b3f7a-913c-43cd-bf74-a5b4df81567c") : error deleting /var/lib/kubelet/pods/0e9b3f7a-913c-43cd-bf74-a5b4df81567c/volume-subpaths: remove /var/lib/kubelet/pods/0e9b3f7a-913c-43cd-bf74-a5b4df81567c/volume-subpaths: no such file or directory Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.192163 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-config" (OuterVolumeSpecName: "config") pod "0e9b3f7a-913c-43cd-bf74-a5b4df81567c" (UID: "0e9b3f7a-913c-43cd-bf74-a5b4df81567c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.232564 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-s87wf"] Feb 14 11:25:28 crc kubenswrapper[4904]: W0214 11:25:28.235810 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod692d42dd_2dae_4d7f_bd38_08720eb781bc.slice/crio-e8c5198e0fb9c7cda54d5ca43ae9dc507dd6788d08f261df4bbb25af661b36c7 WatchSource:0}: Error finding container e8c5198e0fb9c7cda54d5ca43ae9dc507dd6788d08f261df4bbb25af661b36c7: Status 404 returned error can't find the container with id e8c5198e0fb9c7cda54d5ca43ae9dc507dd6788d08f261df4bbb25af661b36c7 Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.247627 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v99fl\" (UniqueName: \"kubernetes.io/projected/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-kube-api-access-v99fl\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.247653 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.393090 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 14 11:25:28 crc kubenswrapper[4904]: W0214 11:25:28.394403 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeed3ea68_257f_49bd_8a2b_4f86a34590c9.slice/crio-45f9fdf18c8d6f9f0befea614c2ba103c8af0da223cbb0dab20121e4b13b3453 WatchSource:0}: Error finding container 45f9fdf18c8d6f9f0befea614c2ba103c8af0da223cbb0dab20121e4b13b3453: Status 404 returned error can't find the container with id 45f9fdf18c8d6f9f0befea614c2ba103c8af0da223cbb0dab20121e4b13b3453 Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.431364 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-2db8z"] Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.431511 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.431592 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.754485 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-dns-svc\") pod \"0e9b3f7a-913c-43cd-bf74-a5b4df81567c\" (UID: \"0e9b3f7a-913c-43cd-bf74-a5b4df81567c\") " Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.755121 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0e9b3f7a-913c-43cd-bf74-a5b4df81567c" (UID: "0e9b3f7a-913c-43cd-bf74-a5b4df81567c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.755233 4904 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e9b3f7a-913c-43cd-bf74-a5b4df81567c-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.803321 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-s87wf" event={"ID":"692d42dd-2dae-4d7f-bd38-08720eb781bc","Type":"ContainerStarted","Data":"3b884162244feacae34b98f363401e5ec02ea308fd83448ed2b549d02ac4a529"} Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.803383 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-s87wf" event={"ID":"692d42dd-2dae-4d7f-bd38-08720eb781bc","Type":"ContainerStarted","Data":"e8c5198e0fb9c7cda54d5ca43ae9dc507dd6788d08f261df4bbb25af661b36c7"} Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.805410 4904 generic.go:334] "Generic (PLEG): container finished" podID="3e75fcff-e55b-43cd-97fb-2c1eeb565168" containerID="6425d2934ebfb3b6a072e6d26f47f7366548c7275765f0140d0b32f348e763fd" exitCode=0 Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.805495 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" event={"ID":"3e75fcff-e55b-43cd-97fb-2c1eeb565168","Type":"ContainerDied","Data":"6425d2934ebfb3b6a072e6d26f47f7366548c7275765f0140d0b32f348e763fd"} Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.805542 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" event={"ID":"3e75fcff-e55b-43cd-97fb-2c1eeb565168","Type":"ContainerStarted","Data":"77f34272a1aff7f268b60adb1117b79a8c5b8b0ff0db8ccfece3769e8c71db79"} Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.807412 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"eed3ea68-257f-49bd-8a2b-4f86a34590c9","Type":"ContainerStarted","Data":"45f9fdf18c8d6f9f0befea614c2ba103c8af0da223cbb0dab20121e4b13b3453"} Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.813778 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" event={"ID":"0e9b3f7a-913c-43cd-bf74-a5b4df81567c","Type":"ContainerDied","Data":"34a29d65190696d9cd41c56e03689ca50d1c4be9a071dc708aa3a0883caee803"} Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.813849 4904 scope.go:117] "RemoveContainer" containerID="2cc7c942f5624ea094277e37a3898d5df3f8e3712d171533a7c2d4b5e060dee9" Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.813799 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qhdgb" Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.822701 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-s87wf" podStartSLOduration=1.822658916 podStartE2EDuration="1.822658916s" podCreationTimestamp="2026-02-14 11:25:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:25:28.821805213 +0000 UTC m=+919.634569864" watchObservedRunningTime="2026-02-14 11:25:28.822658916 +0000 UTC m=+919.635423577" Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.845454 4904 generic.go:334] "Generic (PLEG): container finished" podID="b6914d96-01d9-45d6-933a-b3b8875349df" containerID="0bd481385927297abfca39eec4b813ec7ff29a4916d09495dfc078175e33205e" exitCode=0 Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.847684 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" event={"ID":"b6914d96-01d9-45d6-933a-b3b8875349df","Type":"ContainerDied","Data":"0bd481385927297abfca39eec4b813ec7ff29a4916d09495dfc078175e33205e"} Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.847984 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" event={"ID":"b6914d96-01d9-45d6-933a-b3b8875349df","Type":"ContainerStarted","Data":"03e9ed03161d94a2f55f24484963ec30545014ca134be8835627d0dd67221e7f"} Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.988923 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qhdgb"] Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.998639 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qhdgb"] Feb 14 11:25:28 crc kubenswrapper[4904]: I0214 11:25:28.998861 4904 scope.go:117] "RemoveContainer" containerID="64b1382e277f66b1de49cccc115892dd00d8679ffa470cb5ec953346a505cf0c" Feb 14 11:25:29 crc kubenswrapper[4904]: I0214 11:25:29.578048 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Feb 14 11:25:29 crc kubenswrapper[4904]: I0214 11:25:29.578079 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Feb 14 11:25:29 crc kubenswrapper[4904]: I0214 11:25:29.866966 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e9b3f7a-913c-43cd-bf74-a5b4df81567c" path="/var/lib/kubelet/pods/0e9b3f7a-913c-43cd-bf74-a5b4df81567c/volumes" Feb 14 11:25:29 crc kubenswrapper[4904]: I0214 11:25:29.867698 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28504613-d82c-4903-a641-684ab0e0c4c3" path="/var/lib/kubelet/pods/28504613-d82c-4903-a641-684ab0e0c4c3/volumes" Feb 14 11:25:29 crc kubenswrapper[4904]: I0214 11:25:29.871826 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" event={"ID":"b6914d96-01d9-45d6-933a-b3b8875349df","Type":"ContainerStarted","Data":"9db01e99a8ed41c20fe7746a677b1d37e6c41f5906ede9ab77f87d7f9b1b3a92"} Feb 14 11:25:29 crc kubenswrapper[4904]: I0214 11:25:29.871945 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:29 crc kubenswrapper[4904]: I0214 11:25:29.875024 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" event={"ID":"3e75fcff-e55b-43cd-97fb-2c1eeb565168","Type":"ContainerStarted","Data":"af09c8fd00106b22dbc64a98c09fd6c660eb330549fc9fa3f26a1dcedc14b0a5"} Feb 14 11:25:29 crc kubenswrapper[4904]: I0214 11:25:29.875297 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" Feb 14 11:25:29 crc kubenswrapper[4904]: I0214 11:25:29.881682 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"eed3ea68-257f-49bd-8a2b-4f86a34590c9","Type":"ContainerStarted","Data":"ce6f6dd6a62b3e6b80ecc43e9b259a101fc9379caa802b45961047b8072d451f"} Feb 14 11:25:29 crc kubenswrapper[4904]: I0214 11:25:29.881732 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"eed3ea68-257f-49bd-8a2b-4f86a34590c9","Type":"ContainerStarted","Data":"c2fe490cd4d0d154980af1d52990378d1302a5364417ce9f95cfb087453e7485"} Feb 14 11:25:29 crc kubenswrapper[4904]: I0214 11:25:29.899290 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" podStartSLOduration=3.899273655 podStartE2EDuration="3.899273655s" podCreationTimestamp="2026-02-14 11:25:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:25:29.896971882 +0000 UTC m=+920.709736543" watchObservedRunningTime="2026-02-14 11:25:29.899273655 +0000 UTC m=+920.712038316" Feb 14 11:25:29 crc kubenswrapper[4904]: I0214 11:25:29.915385 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" podStartSLOduration=2.915370736 podStartE2EDuration="2.915370736s" podCreationTimestamp="2026-02-14 11:25:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:25:29.912566089 +0000 UTC m=+920.725330750" watchObservedRunningTime="2026-02-14 11:25:29.915370736 +0000 UTC m=+920.728135397" Feb 14 11:25:29 crc kubenswrapper[4904]: I0214 11:25:29.953330 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Feb 14 11:25:30 crc kubenswrapper[4904]: I0214 11:25:30.886794 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Feb 14 11:25:30 crc kubenswrapper[4904]: I0214 11:25:30.911417 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.830806246 podStartE2EDuration="3.911400855s" podCreationTimestamp="2026-02-14 11:25:27 +0000 UTC" firstStartedPulling="2026-02-14 11:25:28.395982887 +0000 UTC m=+919.208747548" lastFinishedPulling="2026-02-14 11:25:29.476577496 +0000 UTC m=+920.289342157" observedRunningTime="2026-02-14 11:25:30.905828833 +0000 UTC m=+921.718593494" watchObservedRunningTime="2026-02-14 11:25:30.911400855 +0000 UTC m=+921.724165516" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.366368 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-5vxdt"] Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.366916 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" podUID="3e75fcff-e55b-43cd-97fb-2c1eeb565168" containerName="dnsmasq-dns" containerID="cri-o://af09c8fd00106b22dbc64a98c09fd6c660eb330549fc9fa3f26a1dcedc14b0a5" gracePeriod=10 Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.432416 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-dp9lt"] Feb 14 11:25:32 crc kubenswrapper[4904]: E0214 11:25:32.432698 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28504613-d82c-4903-a641-684ab0e0c4c3" containerName="init" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.432714 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="28504613-d82c-4903-a641-684ab0e0c4c3" containerName="init" Feb 14 11:25:32 crc kubenswrapper[4904]: E0214 11:25:32.432735 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e9b3f7a-913c-43cd-bf74-a5b4df81567c" containerName="init" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.432742 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e9b3f7a-913c-43cd-bf74-a5b4df81567c" containerName="init" Feb 14 11:25:32 crc kubenswrapper[4904]: E0214 11:25:32.432760 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28504613-d82c-4903-a641-684ab0e0c4c3" containerName="dnsmasq-dns" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.432767 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="28504613-d82c-4903-a641-684ab0e0c4c3" containerName="dnsmasq-dns" Feb 14 11:25:32 crc kubenswrapper[4904]: E0214 11:25:32.432790 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e9b3f7a-913c-43cd-bf74-a5b4df81567c" containerName="dnsmasq-dns" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.432795 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e9b3f7a-913c-43cd-bf74-a5b4df81567c" containerName="dnsmasq-dns" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.432968 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e9b3f7a-913c-43cd-bf74-a5b4df81567c" containerName="dnsmasq-dns" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.432980 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="28504613-d82c-4903-a641-684ab0e0c4c3" containerName="dnsmasq-dns" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.433753 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.511978 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-dns-svc\") pod \"dnsmasq-dns-698758b865-dp9lt\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.512039 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znh26\" (UniqueName: \"kubernetes.io/projected/4d6d0a8f-2600-47c4-95a0-10f308e493b5-kube-api-access-znh26\") pod \"dnsmasq-dns-698758b865-dp9lt\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.512098 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-dp9lt\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.512116 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-dp9lt\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.512157 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-config\") pod \"dnsmasq-dns-698758b865-dp9lt\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.514296 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-dp9lt"] Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.575570 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.613101 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-dp9lt\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.613148 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-dp9lt\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.613208 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-config\") pod \"dnsmasq-dns-698758b865-dp9lt\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.613265 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-dns-svc\") pod \"dnsmasq-dns-698758b865-dp9lt\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.613304 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znh26\" (UniqueName: \"kubernetes.io/projected/4d6d0a8f-2600-47c4-95a0-10f308e493b5-kube-api-access-znh26\") pod \"dnsmasq-dns-698758b865-dp9lt\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.614169 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-dp9lt\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.614401 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-dp9lt\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.614739 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-config\") pod \"dnsmasq-dns-698758b865-dp9lt\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.616462 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-dns-svc\") pod \"dnsmasq-dns-698758b865-dp9lt\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.635638 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znh26\" (UniqueName: \"kubernetes.io/projected/4d6d0a8f-2600-47c4-95a0-10f308e493b5-kube-api-access-znh26\") pod \"dnsmasq-dns-698758b865-dp9lt\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.692118 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.752725 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.909393 4904 generic.go:334] "Generic (PLEG): container finished" podID="3e75fcff-e55b-43cd-97fb-2c1eeb565168" containerID="af09c8fd00106b22dbc64a98c09fd6c660eb330549fc9fa3f26a1dcedc14b0a5" exitCode=0 Feb 14 11:25:32 crc kubenswrapper[4904]: I0214 11:25:32.909421 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" event={"ID":"3e75fcff-e55b-43cd-97fb-2c1eeb565168","Type":"ContainerDied","Data":"af09c8fd00106b22dbc64a98c09fd6c660eb330549fc9fa3f26a1dcedc14b0a5"} Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.230431 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-dp9lt"] Feb 14 11:25:33 crc kubenswrapper[4904]: W0214 11:25:33.256171 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d6d0a8f_2600_47c4_95a0_10f308e493b5.slice/crio-3304e565c579de61731cc043b8b0be89bf7ac35207157a022d9041a6f8f4cb4b WatchSource:0}: Error finding container 3304e565c579de61731cc043b8b0be89bf7ac35207157a022d9041a6f8f4cb4b: Status 404 returned error can't find the container with id 3304e565c579de61731cc043b8b0be89bf7ac35207157a022d9041a6f8f4cb4b Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.283989 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.430714 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e75fcff-e55b-43cd-97fb-2c1eeb565168-dns-svc\") pod \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\" (UID: \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\") " Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.431003 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jq8pg\" (UniqueName: \"kubernetes.io/projected/3e75fcff-e55b-43cd-97fb-2c1eeb565168-kube-api-access-jq8pg\") pod \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\" (UID: \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\") " Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.431076 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e75fcff-e55b-43cd-97fb-2c1eeb565168-config\") pod \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\" (UID: \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\") " Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.431096 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e75fcff-e55b-43cd-97fb-2c1eeb565168-ovsdbserver-sb\") pod \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\" (UID: \"3e75fcff-e55b-43cd-97fb-2c1eeb565168\") " Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.438099 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e75fcff-e55b-43cd-97fb-2c1eeb565168-kube-api-access-jq8pg" (OuterVolumeSpecName: "kube-api-access-jq8pg") pod "3e75fcff-e55b-43cd-97fb-2c1eeb565168" (UID: "3e75fcff-e55b-43cd-97fb-2c1eeb565168"). InnerVolumeSpecName "kube-api-access-jq8pg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.467402 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e75fcff-e55b-43cd-97fb-2c1eeb565168-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3e75fcff-e55b-43cd-97fb-2c1eeb565168" (UID: "3e75fcff-e55b-43cd-97fb-2c1eeb565168"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.471434 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e75fcff-e55b-43cd-97fb-2c1eeb565168-config" (OuterVolumeSpecName: "config") pod "3e75fcff-e55b-43cd-97fb-2c1eeb565168" (UID: "3e75fcff-e55b-43cd-97fb-2c1eeb565168"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.473374 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e75fcff-e55b-43cd-97fb-2c1eeb565168-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3e75fcff-e55b-43cd-97fb-2c1eeb565168" (UID: "3e75fcff-e55b-43cd-97fb-2c1eeb565168"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.533435 4904 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e75fcff-e55b-43cd-97fb-2c1eeb565168-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.533469 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jq8pg\" (UniqueName: \"kubernetes.io/projected/3e75fcff-e55b-43cd-97fb-2c1eeb565168-kube-api-access-jq8pg\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.533484 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e75fcff-e55b-43cd-97fb-2c1eeb565168-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.533493 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e75fcff-e55b-43cd-97fb-2c1eeb565168-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.571161 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Feb 14 11:25:33 crc kubenswrapper[4904]: E0214 11:25:33.571629 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e75fcff-e55b-43cd-97fb-2c1eeb565168" containerName="dnsmasq-dns" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.571694 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e75fcff-e55b-43cd-97fb-2c1eeb565168" containerName="dnsmasq-dns" Feb 14 11:25:33 crc kubenswrapper[4904]: E0214 11:25:33.571811 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e75fcff-e55b-43cd-97fb-2c1eeb565168" containerName="init" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.571883 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e75fcff-e55b-43cd-97fb-2c1eeb565168" containerName="init" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.572101 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e75fcff-e55b-43cd-97fb-2c1eeb565168" containerName="dnsmasq-dns" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.576601 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.578959 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.578972 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.579875 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-4rdsn" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.580127 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.595945 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.736363 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0594337d-464f-489f-9761-3d3d2e143af8-lock\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.736412 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.736438 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.736471 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0594337d-464f-489f-9761-3d3d2e143af8-cache\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.736539 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0594337d-464f-489f-9761-3d3d2e143af8-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.736765 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbm7s\" (UniqueName: \"kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-kube-api-access-lbm7s\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.838169 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0594337d-464f-489f-9761-3d3d2e143af8-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.838247 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbm7s\" (UniqueName: \"kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-kube-api-access-lbm7s\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.838272 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0594337d-464f-489f-9761-3d3d2e143af8-lock\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.838292 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.838311 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.838327 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0594337d-464f-489f-9761-3d3d2e143af8-cache\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:33 crc kubenswrapper[4904]: E0214 11:25:33.838499 4904 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 14 11:25:33 crc kubenswrapper[4904]: E0214 11:25:33.838531 4904 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 14 11:25:33 crc kubenswrapper[4904]: E0214 11:25:33.838596 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift podName:0594337d-464f-489f-9761-3d3d2e143af8 nodeName:}" failed. No retries permitted until 2026-02-14 11:25:34.338575193 +0000 UTC m=+925.151339854 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift") pod "swift-storage-0" (UID: "0594337d-464f-489f-9761-3d3d2e143af8") : configmap "swift-ring-files" not found Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.838600 4904 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/swift-storage-0" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.838776 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0594337d-464f-489f-9761-3d3d2e143af8-lock\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.839088 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0594337d-464f-489f-9761-3d3d2e143af8-cache\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.855255 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0594337d-464f-489f-9761-3d3d2e143af8-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.871152 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.875259 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbm7s\" (UniqueName: \"kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-kube-api-access-lbm7s\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.916267 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.916739 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-5vxdt" event={"ID":"3e75fcff-e55b-43cd-97fb-2c1eeb565168","Type":"ContainerDied","Data":"77f34272a1aff7f268b60adb1117b79a8c5b8b0ff0db8ccfece3769e8c71db79"} Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.916818 4904 scope.go:117] "RemoveContainer" containerID="af09c8fd00106b22dbc64a98c09fd6c660eb330549fc9fa3f26a1dcedc14b0a5" Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.918470 4904 generic.go:334] "Generic (PLEG): container finished" podID="4d6d0a8f-2600-47c4-95a0-10f308e493b5" containerID="5a3f35624c4a676928fb933326f2b0137135f8c9e439ee555d4fcd7f836c9026" exitCode=0 Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.918513 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-dp9lt" event={"ID":"4d6d0a8f-2600-47c4-95a0-10f308e493b5","Type":"ContainerDied","Data":"5a3f35624c4a676928fb933326f2b0137135f8c9e439ee555d4fcd7f836c9026"} Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.918538 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-dp9lt" event={"ID":"4d6d0a8f-2600-47c4-95a0-10f308e493b5","Type":"ContainerStarted","Data":"3304e565c579de61731cc043b8b0be89bf7ac35207157a022d9041a6f8f4cb4b"} Feb 14 11:25:33 crc kubenswrapper[4904]: I0214 11:25:33.938530 4904 scope.go:117] "RemoveContainer" containerID="6425d2934ebfb3b6a072e6d26f47f7366548c7275765f0140d0b32f348e763fd" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.044621 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-5vxdt"] Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.046274 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-5vxdt"] Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.097731 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-jpt8m"] Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.098788 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.103293 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.103457 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.103629 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.113077 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-jpt8m"] Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.134248 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-tw4ls"] Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.135588 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.157348 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-jpt8m"] Feb 14 11:25:34 crc kubenswrapper[4904]: E0214 11:25:34.157870 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-5mcdf ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-5mcdf ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-jpt8m" podUID="d8a55dbd-1676-4b64-8ebe-ccb6c1a34996" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.168113 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-tw4ls"] Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.190882 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.250546 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-combined-ca-bundle\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.250611 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-swiftconf\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.250656 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2a79513-6beb-47b8-96e5-b5424056661e-ring-data-devices\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.250683 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s9dz\" (UniqueName: \"kubernetes.io/projected/e2a79513-6beb-47b8-96e5-b5424056661e-kube-api-access-2s9dz\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.250723 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-dispersionconf\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.250743 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-scripts\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.250794 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2a79513-6beb-47b8-96e5-b5424056661e-swiftconf\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.250819 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2a79513-6beb-47b8-96e5-b5424056661e-scripts\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.250877 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2a79513-6beb-47b8-96e5-b5424056661e-dispersionconf\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.250902 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-ring-data-devices\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.250933 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2a79513-6beb-47b8-96e5-b5424056661e-combined-ca-bundle\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.250956 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mcdf\" (UniqueName: \"kubernetes.io/projected/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-kube-api-access-5mcdf\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.250990 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-etc-swift\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.251015 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2a79513-6beb-47b8-96e5-b5424056661e-etc-swift\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.352280 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-swiftconf\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.352341 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2a79513-6beb-47b8-96e5-b5424056661e-ring-data-devices\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.352361 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s9dz\" (UniqueName: \"kubernetes.io/projected/e2a79513-6beb-47b8-96e5-b5424056661e-kube-api-access-2s9dz\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.352398 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-dispersionconf\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.352413 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-scripts\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.352442 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.352461 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2a79513-6beb-47b8-96e5-b5424056661e-swiftconf\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.352481 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2a79513-6beb-47b8-96e5-b5424056661e-scripts\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.352524 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2a79513-6beb-47b8-96e5-b5424056661e-dispersionconf\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.352542 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-ring-data-devices\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.352564 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2a79513-6beb-47b8-96e5-b5424056661e-combined-ca-bundle\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.352580 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mcdf\" (UniqueName: \"kubernetes.io/projected/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-kube-api-access-5mcdf\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.352607 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-etc-swift\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.352629 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2a79513-6beb-47b8-96e5-b5424056661e-etc-swift\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.352650 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-combined-ca-bundle\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: E0214 11:25:34.353360 4904 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 14 11:25:34 crc kubenswrapper[4904]: E0214 11:25:34.353379 4904 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 14 11:25:34 crc kubenswrapper[4904]: E0214 11:25:34.353422 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift podName:0594337d-464f-489f-9761-3d3d2e143af8 nodeName:}" failed. No retries permitted until 2026-02-14 11:25:35.353406789 +0000 UTC m=+926.166171450 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift") pod "swift-storage-0" (UID: "0594337d-464f-489f-9761-3d3d2e143af8") : configmap "swift-ring-files" not found Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.354264 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2a79513-6beb-47b8-96e5-b5424056661e-ring-data-devices\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.354271 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-ring-data-devices\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.354307 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2a79513-6beb-47b8-96e5-b5424056661e-scripts\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.354509 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-etc-swift\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.355111 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-scripts\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.355363 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2a79513-6beb-47b8-96e5-b5424056661e-etc-swift\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.363246 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-dispersionconf\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.367156 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2a79513-6beb-47b8-96e5-b5424056661e-dispersionconf\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.367323 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2a79513-6beb-47b8-96e5-b5424056661e-swiftconf\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.367593 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2a79513-6beb-47b8-96e5-b5424056661e-combined-ca-bundle\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.367770 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-swiftconf\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.369445 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-combined-ca-bundle\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.376170 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mcdf\" (UniqueName: \"kubernetes.io/projected/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-kube-api-access-5mcdf\") pod \"swift-ring-rebalance-jpt8m\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.381412 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s9dz\" (UniqueName: \"kubernetes.io/projected/e2a79513-6beb-47b8-96e5-b5424056661e-kube-api-access-2s9dz\") pod \"swift-ring-rebalance-tw4ls\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.452786 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.473103 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.916167 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-tw4ls"] Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.935865 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tw4ls" event={"ID":"e2a79513-6beb-47b8-96e5-b5424056661e","Type":"ContainerStarted","Data":"bdb42885c87db0ed09b0d05217ad3125bcdcddb2071442bb14d180181d97d752"} Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.937549 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-dp9lt" event={"ID":"4d6d0a8f-2600-47c4-95a0-10f308e493b5","Type":"ContainerStarted","Data":"68bcf29edaf47bf7889aa5762ca5390fc72fa16206dd5f0776a150a1958a2278"} Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.937702 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.940555 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.962349 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-dp9lt" podStartSLOduration=2.962332605 podStartE2EDuration="2.962332605s" podCreationTimestamp="2026-02-14 11:25:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:25:34.956117564 +0000 UTC m=+925.768882225" watchObservedRunningTime="2026-02-14 11:25:34.962332605 +0000 UTC m=+925.775097266" Feb 14 11:25:34 crc kubenswrapper[4904]: I0214 11:25:34.967819 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.061604 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-swiftconf\") pod \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.061705 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-combined-ca-bundle\") pod \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.061764 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-dispersionconf\") pod \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.061800 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mcdf\" (UniqueName: \"kubernetes.io/projected/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-kube-api-access-5mcdf\") pod \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.061878 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-ring-data-devices\") pod \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.061901 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-scripts\") pod \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.061919 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-etc-swift\") pod \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\" (UID: \"d8a55dbd-1676-4b64-8ebe-ccb6c1a34996\") " Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.063144 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d8a55dbd-1676-4b64-8ebe-ccb6c1a34996" (UID: "d8a55dbd-1676-4b64-8ebe-ccb6c1a34996"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.063378 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-scripts" (OuterVolumeSpecName: "scripts") pod "d8a55dbd-1676-4b64-8ebe-ccb6c1a34996" (UID: "d8a55dbd-1676-4b64-8ebe-ccb6c1a34996"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.063490 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d8a55dbd-1676-4b64-8ebe-ccb6c1a34996" (UID: "d8a55dbd-1676-4b64-8ebe-ccb6c1a34996"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.069111 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d8a55dbd-1676-4b64-8ebe-ccb6c1a34996" (UID: "d8a55dbd-1676-4b64-8ebe-ccb6c1a34996"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.069135 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-kube-api-access-5mcdf" (OuterVolumeSpecName: "kube-api-access-5mcdf") pod "d8a55dbd-1676-4b64-8ebe-ccb6c1a34996" (UID: "d8a55dbd-1676-4b64-8ebe-ccb6c1a34996"). InnerVolumeSpecName "kube-api-access-5mcdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.070334 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d8a55dbd-1676-4b64-8ebe-ccb6c1a34996" (UID: "d8a55dbd-1676-4b64-8ebe-ccb6c1a34996"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.076405 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8a55dbd-1676-4b64-8ebe-ccb6c1a34996" (UID: "d8a55dbd-1676-4b64-8ebe-ccb6c1a34996"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.163456 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.163483 4904 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.163493 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mcdf\" (UniqueName: \"kubernetes.io/projected/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-kube-api-access-5mcdf\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.163503 4904 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.163515 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.163527 4904 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.163537 4904 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.367277 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:35 crc kubenswrapper[4904]: E0214 11:25:35.367434 4904 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 14 11:25:35 crc kubenswrapper[4904]: E0214 11:25:35.367448 4904 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 14 11:25:35 crc kubenswrapper[4904]: E0214 11:25:35.367495 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift podName:0594337d-464f-489f-9761-3d3d2e143af8 nodeName:}" failed. No retries permitted until 2026-02-14 11:25:37.367481793 +0000 UTC m=+928.180246444 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift") pod "swift-storage-0" (UID: "0594337d-464f-489f-9761-3d3d2e143af8") : configmap "swift-ring-files" not found Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.870578 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e75fcff-e55b-43cd-97fb-2c1eeb565168" path="/var/lib/kubelet/pods/3e75fcff-e55b-43cd-97fb-2c1eeb565168/volumes" Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.946779 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jpt8m" Feb 14 11:25:35 crc kubenswrapper[4904]: I0214 11:25:35.993557 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-jpt8m"] Feb 14 11:25:36 crc kubenswrapper[4904]: I0214 11:25:36.000676 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-jpt8m"] Feb 14 11:25:36 crc kubenswrapper[4904]: I0214 11:25:36.222203 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wcd5n" Feb 14 11:25:36 crc kubenswrapper[4904]: I0214 11:25:36.280771 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wcd5n" Feb 14 11:25:36 crc kubenswrapper[4904]: I0214 11:25:36.487670 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wcd5n"] Feb 14 11:25:37 crc kubenswrapper[4904]: I0214 11:25:37.112252 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-s9htl"] Feb 14 11:25:37 crc kubenswrapper[4904]: I0214 11:25:37.116385 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-s9htl" Feb 14 11:25:37 crc kubenswrapper[4904]: I0214 11:25:37.118981 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 14 11:25:37 crc kubenswrapper[4904]: I0214 11:25:37.121325 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-s9htl"] Feb 14 11:25:37 crc kubenswrapper[4904]: I0214 11:25:37.211904 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f52c7efb-196a-4ced-aaba-a380e99098ff-operator-scripts\") pod \"root-account-create-update-s9htl\" (UID: \"f52c7efb-196a-4ced-aaba-a380e99098ff\") " pod="openstack/root-account-create-update-s9htl" Feb 14 11:25:37 crc kubenswrapper[4904]: I0214 11:25:37.212187 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsvgc\" (UniqueName: \"kubernetes.io/projected/f52c7efb-196a-4ced-aaba-a380e99098ff-kube-api-access-nsvgc\") pod \"root-account-create-update-s9htl\" (UID: \"f52c7efb-196a-4ced-aaba-a380e99098ff\") " pod="openstack/root-account-create-update-s9htl" Feb 14 11:25:37 crc kubenswrapper[4904]: I0214 11:25:37.313946 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f52c7efb-196a-4ced-aaba-a380e99098ff-operator-scripts\") pod \"root-account-create-update-s9htl\" (UID: \"f52c7efb-196a-4ced-aaba-a380e99098ff\") " pod="openstack/root-account-create-update-s9htl" Feb 14 11:25:37 crc kubenswrapper[4904]: I0214 11:25:37.314018 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsvgc\" (UniqueName: \"kubernetes.io/projected/f52c7efb-196a-4ced-aaba-a380e99098ff-kube-api-access-nsvgc\") pod \"root-account-create-update-s9htl\" (UID: \"f52c7efb-196a-4ced-aaba-a380e99098ff\") " pod="openstack/root-account-create-update-s9htl" Feb 14 11:25:37 crc kubenswrapper[4904]: I0214 11:25:37.315808 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f52c7efb-196a-4ced-aaba-a380e99098ff-operator-scripts\") pod \"root-account-create-update-s9htl\" (UID: \"f52c7efb-196a-4ced-aaba-a380e99098ff\") " pod="openstack/root-account-create-update-s9htl" Feb 14 11:25:37 crc kubenswrapper[4904]: I0214 11:25:37.352794 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsvgc\" (UniqueName: \"kubernetes.io/projected/f52c7efb-196a-4ced-aaba-a380e99098ff-kube-api-access-nsvgc\") pod \"root-account-create-update-s9htl\" (UID: \"f52c7efb-196a-4ced-aaba-a380e99098ff\") " pod="openstack/root-account-create-update-s9htl" Feb 14 11:25:37 crc kubenswrapper[4904]: I0214 11:25:37.415685 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:37 crc kubenswrapper[4904]: E0214 11:25:37.415904 4904 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 14 11:25:37 crc kubenswrapper[4904]: E0214 11:25:37.415956 4904 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 14 11:25:37 crc kubenswrapper[4904]: E0214 11:25:37.416017 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift podName:0594337d-464f-489f-9761-3d3d2e143af8 nodeName:}" failed. No retries permitted until 2026-02-14 11:25:41.41600187 +0000 UTC m=+932.228766531 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift") pod "swift-storage-0" (UID: "0594337d-464f-489f-9761-3d3d2e143af8") : configmap "swift-ring-files" not found Feb 14 11:25:37 crc kubenswrapper[4904]: I0214 11:25:37.446191 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-s9htl" Feb 14 11:25:37 crc kubenswrapper[4904]: I0214 11:25:37.779102 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:37 crc kubenswrapper[4904]: I0214 11:25:37.869283 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8a55dbd-1676-4b64-8ebe-ccb6c1a34996" path="/var/lib/kubelet/pods/d8a55dbd-1676-4b64-8ebe-ccb6c1a34996/volumes" Feb 14 11:25:37 crc kubenswrapper[4904]: I0214 11:25:37.965690 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wcd5n" podUID="28264935-4532-4d4e-b795-a8b927395d2e" containerName="registry-server" containerID="cri-o://067ab3afa3a539abf7f0361c03fe3e66c38b9aba0764cdaa8573cd07f2b5994f" gracePeriod=2 Feb 14 11:25:38 crc kubenswrapper[4904]: I0214 11:25:38.706324 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wcd5n" Feb 14 11:25:38 crc kubenswrapper[4904]: I0214 11:25:38.848582 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28264935-4532-4d4e-b795-a8b927395d2e-utilities\") pod \"28264935-4532-4d4e-b795-a8b927395d2e\" (UID: \"28264935-4532-4d4e-b795-a8b927395d2e\") " Feb 14 11:25:38 crc kubenswrapper[4904]: I0214 11:25:38.848728 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28264935-4532-4d4e-b795-a8b927395d2e-catalog-content\") pod \"28264935-4532-4d4e-b795-a8b927395d2e\" (UID: \"28264935-4532-4d4e-b795-a8b927395d2e\") " Feb 14 11:25:38 crc kubenswrapper[4904]: I0214 11:25:38.848828 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmxgv\" (UniqueName: \"kubernetes.io/projected/28264935-4532-4d4e-b795-a8b927395d2e-kube-api-access-qmxgv\") pod \"28264935-4532-4d4e-b795-a8b927395d2e\" (UID: \"28264935-4532-4d4e-b795-a8b927395d2e\") " Feb 14 11:25:38 crc kubenswrapper[4904]: I0214 11:25:38.849446 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28264935-4532-4d4e-b795-a8b927395d2e-utilities" (OuterVolumeSpecName: "utilities") pod "28264935-4532-4d4e-b795-a8b927395d2e" (UID: "28264935-4532-4d4e-b795-a8b927395d2e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:25:38 crc kubenswrapper[4904]: I0214 11:25:38.854258 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28264935-4532-4d4e-b795-a8b927395d2e-kube-api-access-qmxgv" (OuterVolumeSpecName: "kube-api-access-qmxgv") pod "28264935-4532-4d4e-b795-a8b927395d2e" (UID: "28264935-4532-4d4e-b795-a8b927395d2e"). InnerVolumeSpecName "kube-api-access-qmxgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:25:38 crc kubenswrapper[4904]: I0214 11:25:38.894804 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28264935-4532-4d4e-b795-a8b927395d2e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "28264935-4532-4d4e-b795-a8b927395d2e" (UID: "28264935-4532-4d4e-b795-a8b927395d2e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:25:38 crc kubenswrapper[4904]: I0214 11:25:38.931547 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-s9htl"] Feb 14 11:25:38 crc kubenswrapper[4904]: W0214 11:25:38.940627 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf52c7efb_196a_4ced_aaba_a380e99098ff.slice/crio-a9ef4420644c30eb6ddd887e6c73b6f1495b8ab307cd9f58d4e55d79360f622f WatchSource:0}: Error finding container a9ef4420644c30eb6ddd887e6c73b6f1495b8ab307cd9f58d4e55d79360f622f: Status 404 returned error can't find the container with id a9ef4420644c30eb6ddd887e6c73b6f1495b8ab307cd9f58d4e55d79360f622f Feb 14 11:25:38 crc kubenswrapper[4904]: I0214 11:25:38.950959 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmxgv\" (UniqueName: \"kubernetes.io/projected/28264935-4532-4d4e-b795-a8b927395d2e-kube-api-access-qmxgv\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:38 crc kubenswrapper[4904]: I0214 11:25:38.951090 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28264935-4532-4d4e-b795-a8b927395d2e-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:38 crc kubenswrapper[4904]: I0214 11:25:38.951168 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28264935-4532-4d4e-b795-a8b927395d2e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:38 crc kubenswrapper[4904]: I0214 11:25:38.973437 4904 generic.go:334] "Generic (PLEG): container finished" podID="28264935-4532-4d4e-b795-a8b927395d2e" containerID="067ab3afa3a539abf7f0361c03fe3e66c38b9aba0764cdaa8573cd07f2b5994f" exitCode=0 Feb 14 11:25:38 crc kubenswrapper[4904]: I0214 11:25:38.973695 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wcd5n" Feb 14 11:25:38 crc kubenswrapper[4904]: I0214 11:25:38.973588 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wcd5n" event={"ID":"28264935-4532-4d4e-b795-a8b927395d2e","Type":"ContainerDied","Data":"067ab3afa3a539abf7f0361c03fe3e66c38b9aba0764cdaa8573cd07f2b5994f"} Feb 14 11:25:38 crc kubenswrapper[4904]: I0214 11:25:38.974044 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wcd5n" event={"ID":"28264935-4532-4d4e-b795-a8b927395d2e","Type":"ContainerDied","Data":"d72b16ce3deb53adfb9e7dbdb8e5e8fd7b5e77858a17f4950cd3a121e992e618"} Feb 14 11:25:38 crc kubenswrapper[4904]: I0214 11:25:38.974129 4904 scope.go:117] "RemoveContainer" containerID="067ab3afa3a539abf7f0361c03fe3e66c38b9aba0764cdaa8573cd07f2b5994f" Feb 14 11:25:38 crc kubenswrapper[4904]: I0214 11:25:38.976164 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-s9htl" event={"ID":"f52c7efb-196a-4ced-aaba-a380e99098ff","Type":"ContainerStarted","Data":"a9ef4420644c30eb6ddd887e6c73b6f1495b8ab307cd9f58d4e55d79360f622f"} Feb 14 11:25:38 crc kubenswrapper[4904]: I0214 11:25:38.985910 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tw4ls" event={"ID":"e2a79513-6beb-47b8-96e5-b5424056661e","Type":"ContainerStarted","Data":"a99499067fec11baa46e626f0ebc36bea23368382a810a4827c0d937c2719e48"} Feb 14 11:25:39 crc kubenswrapper[4904]: I0214 11:25:39.003955 4904 scope.go:117] "RemoveContainer" containerID="eb833188011ca621a3318aa7c5b10a299e65eb1586a7f5f5fe9af7d6f09cc32c" Feb 14 11:25:39 crc kubenswrapper[4904]: I0214 11:25:39.011621 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-tw4ls" podStartSLOduration=1.488953154 podStartE2EDuration="5.011603079s" podCreationTimestamp="2026-02-14 11:25:34 +0000 UTC" firstStartedPulling="2026-02-14 11:25:34.92242429 +0000 UTC m=+925.735188951" lastFinishedPulling="2026-02-14 11:25:38.445074225 +0000 UTC m=+929.257838876" observedRunningTime="2026-02-14 11:25:39.00546092 +0000 UTC m=+929.818225581" watchObservedRunningTime="2026-02-14 11:25:39.011603079 +0000 UTC m=+929.824367740" Feb 14 11:25:39 crc kubenswrapper[4904]: I0214 11:25:39.028416 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wcd5n"] Feb 14 11:25:39 crc kubenswrapper[4904]: I0214 11:25:39.047788 4904 scope.go:117] "RemoveContainer" containerID="139aa32e3123828d9af3e7d70b1422d3372f7eef680e428d808c6f4780ae6c75" Feb 14 11:25:39 crc kubenswrapper[4904]: I0214 11:25:39.049923 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wcd5n"] Feb 14 11:25:39 crc kubenswrapper[4904]: I0214 11:25:39.095458 4904 scope.go:117] "RemoveContainer" containerID="067ab3afa3a539abf7f0361c03fe3e66c38b9aba0764cdaa8573cd07f2b5994f" Feb 14 11:25:39 crc kubenswrapper[4904]: E0214 11:25:39.096037 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"067ab3afa3a539abf7f0361c03fe3e66c38b9aba0764cdaa8573cd07f2b5994f\": container with ID starting with 067ab3afa3a539abf7f0361c03fe3e66c38b9aba0764cdaa8573cd07f2b5994f not found: ID does not exist" containerID="067ab3afa3a539abf7f0361c03fe3e66c38b9aba0764cdaa8573cd07f2b5994f" Feb 14 11:25:39 crc kubenswrapper[4904]: I0214 11:25:39.096163 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"067ab3afa3a539abf7f0361c03fe3e66c38b9aba0764cdaa8573cd07f2b5994f"} err="failed to get container status \"067ab3afa3a539abf7f0361c03fe3e66c38b9aba0764cdaa8573cd07f2b5994f\": rpc error: code = NotFound desc = could not find container \"067ab3afa3a539abf7f0361c03fe3e66c38b9aba0764cdaa8573cd07f2b5994f\": container with ID starting with 067ab3afa3a539abf7f0361c03fe3e66c38b9aba0764cdaa8573cd07f2b5994f not found: ID does not exist" Feb 14 11:25:39 crc kubenswrapper[4904]: I0214 11:25:39.096238 4904 scope.go:117] "RemoveContainer" containerID="eb833188011ca621a3318aa7c5b10a299e65eb1586a7f5f5fe9af7d6f09cc32c" Feb 14 11:25:39 crc kubenswrapper[4904]: E0214 11:25:39.096556 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb833188011ca621a3318aa7c5b10a299e65eb1586a7f5f5fe9af7d6f09cc32c\": container with ID starting with eb833188011ca621a3318aa7c5b10a299e65eb1586a7f5f5fe9af7d6f09cc32c not found: ID does not exist" containerID="eb833188011ca621a3318aa7c5b10a299e65eb1586a7f5f5fe9af7d6f09cc32c" Feb 14 11:25:39 crc kubenswrapper[4904]: I0214 11:25:39.096657 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb833188011ca621a3318aa7c5b10a299e65eb1586a7f5f5fe9af7d6f09cc32c"} err="failed to get container status \"eb833188011ca621a3318aa7c5b10a299e65eb1586a7f5f5fe9af7d6f09cc32c\": rpc error: code = NotFound desc = could not find container \"eb833188011ca621a3318aa7c5b10a299e65eb1586a7f5f5fe9af7d6f09cc32c\": container with ID starting with eb833188011ca621a3318aa7c5b10a299e65eb1586a7f5f5fe9af7d6f09cc32c not found: ID does not exist" Feb 14 11:25:39 crc kubenswrapper[4904]: I0214 11:25:39.096733 4904 scope.go:117] "RemoveContainer" containerID="139aa32e3123828d9af3e7d70b1422d3372f7eef680e428d808c6f4780ae6c75" Feb 14 11:25:39 crc kubenswrapper[4904]: E0214 11:25:39.097097 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"139aa32e3123828d9af3e7d70b1422d3372f7eef680e428d808c6f4780ae6c75\": container with ID starting with 139aa32e3123828d9af3e7d70b1422d3372f7eef680e428d808c6f4780ae6c75 not found: ID does not exist" containerID="139aa32e3123828d9af3e7d70b1422d3372f7eef680e428d808c6f4780ae6c75" Feb 14 11:25:39 crc kubenswrapper[4904]: I0214 11:25:39.097175 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"139aa32e3123828d9af3e7d70b1422d3372f7eef680e428d808c6f4780ae6c75"} err="failed to get container status \"139aa32e3123828d9af3e7d70b1422d3372f7eef680e428d808c6f4780ae6c75\": rpc error: code = NotFound desc = could not find container \"139aa32e3123828d9af3e7d70b1422d3372f7eef680e428d808c6f4780ae6c75\": container with ID starting with 139aa32e3123828d9af3e7d70b1422d3372f7eef680e428d808c6f4780ae6c75 not found: ID does not exist" Feb 14 11:25:39 crc kubenswrapper[4904]: I0214 11:25:39.872258 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28264935-4532-4d4e-b795-a8b927395d2e" path="/var/lib/kubelet/pods/28264935-4532-4d4e-b795-a8b927395d2e/volumes" Feb 14 11:25:39 crc kubenswrapper[4904]: I0214 11:25:39.996260 4904 generic.go:334] "Generic (PLEG): container finished" podID="f52c7efb-196a-4ced-aaba-a380e99098ff" containerID="f7b2c2d2959ca38c222d69f4400b8387898daab9f22ebbc77c383e5b4d87cb3b" exitCode=0 Feb 14 11:25:39 crc kubenswrapper[4904]: I0214 11:25:39.996358 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-s9htl" event={"ID":"f52c7efb-196a-4ced-aaba-a380e99098ff","Type":"ContainerDied","Data":"f7b2c2d2959ca38c222d69f4400b8387898daab9f22ebbc77c383e5b4d87cb3b"} Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.098046 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-4dkbs"] Feb 14 11:25:40 crc kubenswrapper[4904]: E0214 11:25:40.098512 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28264935-4532-4d4e-b795-a8b927395d2e" containerName="extract-utilities" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.098536 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="28264935-4532-4d4e-b795-a8b927395d2e" containerName="extract-utilities" Feb 14 11:25:40 crc kubenswrapper[4904]: E0214 11:25:40.098568 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28264935-4532-4d4e-b795-a8b927395d2e" containerName="registry-server" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.098577 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="28264935-4532-4d4e-b795-a8b927395d2e" containerName="registry-server" Feb 14 11:25:40 crc kubenswrapper[4904]: E0214 11:25:40.098597 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28264935-4532-4d4e-b795-a8b927395d2e" containerName="extract-content" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.098605 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="28264935-4532-4d4e-b795-a8b927395d2e" containerName="extract-content" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.098820 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="28264935-4532-4d4e-b795-a8b927395d2e" containerName="registry-server" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.106205 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4dkbs" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.135021 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-4dkbs"] Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.225477 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-981b-account-create-update-rt8h8"] Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.226713 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-981b-account-create-update-rt8h8" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.231032 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.243868 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-981b-account-create-update-rt8h8"] Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.273685 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd9c7534-8ff2-4898-ab20-660140c54857-operator-scripts\") pod \"glance-db-create-4dkbs\" (UID: \"cd9c7534-8ff2-4898-ab20-660140c54857\") " pod="openstack/glance-db-create-4dkbs" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.273743 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-882qx\" (UniqueName: \"kubernetes.io/projected/cd9c7534-8ff2-4898-ab20-660140c54857-kube-api-access-882qx\") pod \"glance-db-create-4dkbs\" (UID: \"cd9c7534-8ff2-4898-ab20-660140c54857\") " pod="openstack/glance-db-create-4dkbs" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.375041 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12e4e520-e219-4768-b5db-5ccadfee8beb-operator-scripts\") pod \"glance-981b-account-create-update-rt8h8\" (UID: \"12e4e520-e219-4768-b5db-5ccadfee8beb\") " pod="openstack/glance-981b-account-create-update-rt8h8" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.375107 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-882qx\" (UniqueName: \"kubernetes.io/projected/cd9c7534-8ff2-4898-ab20-660140c54857-kube-api-access-882qx\") pod \"glance-db-create-4dkbs\" (UID: \"cd9c7534-8ff2-4898-ab20-660140c54857\") " pod="openstack/glance-db-create-4dkbs" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.375204 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqx9c\" (UniqueName: \"kubernetes.io/projected/12e4e520-e219-4768-b5db-5ccadfee8beb-kube-api-access-lqx9c\") pod \"glance-981b-account-create-update-rt8h8\" (UID: \"12e4e520-e219-4768-b5db-5ccadfee8beb\") " pod="openstack/glance-981b-account-create-update-rt8h8" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.375256 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd9c7534-8ff2-4898-ab20-660140c54857-operator-scripts\") pod \"glance-db-create-4dkbs\" (UID: \"cd9c7534-8ff2-4898-ab20-660140c54857\") " pod="openstack/glance-db-create-4dkbs" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.375995 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd9c7534-8ff2-4898-ab20-660140c54857-operator-scripts\") pod \"glance-db-create-4dkbs\" (UID: \"cd9c7534-8ff2-4898-ab20-660140c54857\") " pod="openstack/glance-db-create-4dkbs" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.408414 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-882qx\" (UniqueName: \"kubernetes.io/projected/cd9c7534-8ff2-4898-ab20-660140c54857-kube-api-access-882qx\") pod \"glance-db-create-4dkbs\" (UID: \"cd9c7534-8ff2-4898-ab20-660140c54857\") " pod="openstack/glance-db-create-4dkbs" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.429574 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4dkbs" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.479754 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqx9c\" (UniqueName: \"kubernetes.io/projected/12e4e520-e219-4768-b5db-5ccadfee8beb-kube-api-access-lqx9c\") pod \"glance-981b-account-create-update-rt8h8\" (UID: \"12e4e520-e219-4768-b5db-5ccadfee8beb\") " pod="openstack/glance-981b-account-create-update-rt8h8" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.479871 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12e4e520-e219-4768-b5db-5ccadfee8beb-operator-scripts\") pod \"glance-981b-account-create-update-rt8h8\" (UID: \"12e4e520-e219-4768-b5db-5ccadfee8beb\") " pod="openstack/glance-981b-account-create-update-rt8h8" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.480677 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12e4e520-e219-4768-b5db-5ccadfee8beb-operator-scripts\") pod \"glance-981b-account-create-update-rt8h8\" (UID: \"12e4e520-e219-4768-b5db-5ccadfee8beb\") " pod="openstack/glance-981b-account-create-update-rt8h8" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.500766 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqx9c\" (UniqueName: \"kubernetes.io/projected/12e4e520-e219-4768-b5db-5ccadfee8beb-kube-api-access-lqx9c\") pod \"glance-981b-account-create-update-rt8h8\" (UID: \"12e4e520-e219-4768-b5db-5ccadfee8beb\") " pod="openstack/glance-981b-account-create-update-rt8h8" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.543399 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-981b-account-create-update-rt8h8" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.866906 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-9cs22"] Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.871745 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9cs22" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.877251 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-9cs22"] Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.926654 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-4dkbs"] Feb 14 11:25:40 crc kubenswrapper[4904]: W0214 11:25:40.927027 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd9c7534_8ff2_4898_ab20_660140c54857.slice/crio-29c5cc8b5d2ba8b51f42500299abce21e4c09055c12b6fc8c66889323b9922f6 WatchSource:0}: Error finding container 29c5cc8b5d2ba8b51f42500299abce21e4c09055c12b6fc8c66889323b9922f6: Status 404 returned error can't find the container with id 29c5cc8b5d2ba8b51f42500299abce21e4c09055c12b6fc8c66889323b9922f6 Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.967982 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-836b-account-create-update-kcjv5"] Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.968900 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-836b-account-create-update-kcjv5" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.970973 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.980895 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-836b-account-create-update-kcjv5"] Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.986917 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8649450d-1384-4d27-a4bb-501ebee69f01-operator-scripts\") pod \"keystone-db-create-9cs22\" (UID: \"8649450d-1384-4d27-a4bb-501ebee69f01\") " pod="openstack/keystone-db-create-9cs22" Feb 14 11:25:40 crc kubenswrapper[4904]: I0214 11:25:40.986962 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtp6b\" (UniqueName: \"kubernetes.io/projected/8649450d-1384-4d27-a4bb-501ebee69f01-kube-api-access-xtp6b\") pod \"keystone-db-create-9cs22\" (UID: \"8649450d-1384-4d27-a4bb-501ebee69f01\") " pod="openstack/keystone-db-create-9cs22" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.006308 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4dkbs" event={"ID":"cd9c7534-8ff2-4898-ab20-660140c54857","Type":"ContainerStarted","Data":"29c5cc8b5d2ba8b51f42500299abce21e4c09055c12b6fc8c66889323b9922f6"} Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.028859 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-981b-account-create-update-rt8h8"] Feb 14 11:25:41 crc kubenswrapper[4904]: W0214 11:25:41.039868 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12e4e520_e219_4768_b5db_5ccadfee8beb.slice/crio-f59aba82a53abbb2350b0478a1798271f5c014ea4a97f6dc0e9a2e9306e2c501 WatchSource:0}: Error finding container f59aba82a53abbb2350b0478a1798271f5c014ea4a97f6dc0e9a2e9306e2c501: Status 404 returned error can't find the container with id f59aba82a53abbb2350b0478a1798271f5c014ea4a97f6dc0e9a2e9306e2c501 Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.074985 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-9z2kq"] Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.076252 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-9z2kq" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.087955 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8649450d-1384-4d27-a4bb-501ebee69f01-operator-scripts\") pod \"keystone-db-create-9cs22\" (UID: \"8649450d-1384-4d27-a4bb-501ebee69f01\") " pod="openstack/keystone-db-create-9cs22" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.087995 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtp6b\" (UniqueName: \"kubernetes.io/projected/8649450d-1384-4d27-a4bb-501ebee69f01-kube-api-access-xtp6b\") pod \"keystone-db-create-9cs22\" (UID: \"8649450d-1384-4d27-a4bb-501ebee69f01\") " pod="openstack/keystone-db-create-9cs22" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.088080 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm489\" (UniqueName: \"kubernetes.io/projected/405e6001-9b9c-46bd-8d84-d5005de91285-kube-api-access-lm489\") pod \"keystone-836b-account-create-update-kcjv5\" (UID: \"405e6001-9b9c-46bd-8d84-d5005de91285\") " pod="openstack/keystone-836b-account-create-update-kcjv5" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.088167 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/405e6001-9b9c-46bd-8d84-d5005de91285-operator-scripts\") pod \"keystone-836b-account-create-update-kcjv5\" (UID: \"405e6001-9b9c-46bd-8d84-d5005de91285\") " pod="openstack/keystone-836b-account-create-update-kcjv5" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.088853 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8649450d-1384-4d27-a4bb-501ebee69f01-operator-scripts\") pod \"keystone-db-create-9cs22\" (UID: \"8649450d-1384-4d27-a4bb-501ebee69f01\") " pod="openstack/keystone-db-create-9cs22" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.096050 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-9z2kq"] Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.123042 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtp6b\" (UniqueName: \"kubernetes.io/projected/8649450d-1384-4d27-a4bb-501ebee69f01-kube-api-access-xtp6b\") pod \"keystone-db-create-9cs22\" (UID: \"8649450d-1384-4d27-a4bb-501ebee69f01\") " pod="openstack/keystone-db-create-9cs22" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.172882 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-f0fd-account-create-update-kmmtd"] Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.173882 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f0fd-account-create-update-kmmtd" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.183046 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f0fd-account-create-update-kmmtd"] Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.190177 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cnwk\" (UniqueName: \"kubernetes.io/projected/90945752-c3c4-4aaf-9680-4c5fe89ab1de-kube-api-access-8cnwk\") pod \"placement-db-create-9z2kq\" (UID: \"90945752-c3c4-4aaf-9680-4c5fe89ab1de\") " pod="openstack/placement-db-create-9z2kq" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.190220 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90945752-c3c4-4aaf-9680-4c5fe89ab1de-operator-scripts\") pod \"placement-db-create-9z2kq\" (UID: \"90945752-c3c4-4aaf-9680-4c5fe89ab1de\") " pod="openstack/placement-db-create-9z2kq" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.190256 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm489\" (UniqueName: \"kubernetes.io/projected/405e6001-9b9c-46bd-8d84-d5005de91285-kube-api-access-lm489\") pod \"keystone-836b-account-create-update-kcjv5\" (UID: \"405e6001-9b9c-46bd-8d84-d5005de91285\") " pod="openstack/keystone-836b-account-create-update-kcjv5" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.190337 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/405e6001-9b9c-46bd-8d84-d5005de91285-operator-scripts\") pod \"keystone-836b-account-create-update-kcjv5\" (UID: \"405e6001-9b9c-46bd-8d84-d5005de91285\") " pod="openstack/keystone-836b-account-create-update-kcjv5" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.190932 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/405e6001-9b9c-46bd-8d84-d5005de91285-operator-scripts\") pod \"keystone-836b-account-create-update-kcjv5\" (UID: \"405e6001-9b9c-46bd-8d84-d5005de91285\") " pod="openstack/keystone-836b-account-create-update-kcjv5" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.191052 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9cs22" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.191066 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.218868 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm489\" (UniqueName: \"kubernetes.io/projected/405e6001-9b9c-46bd-8d84-d5005de91285-kube-api-access-lm489\") pod \"keystone-836b-account-create-update-kcjv5\" (UID: \"405e6001-9b9c-46bd-8d84-d5005de91285\") " pod="openstack/keystone-836b-account-create-update-kcjv5" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.304213 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-836b-account-create-update-kcjv5" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.305334 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drv8h\" (UniqueName: \"kubernetes.io/projected/71d5f892-67e3-4854-bee6-1586d712fe8e-kube-api-access-drv8h\") pod \"placement-f0fd-account-create-update-kmmtd\" (UID: \"71d5f892-67e3-4854-bee6-1586d712fe8e\") " pod="openstack/placement-f0fd-account-create-update-kmmtd" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.305375 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cnwk\" (UniqueName: \"kubernetes.io/projected/90945752-c3c4-4aaf-9680-4c5fe89ab1de-kube-api-access-8cnwk\") pod \"placement-db-create-9z2kq\" (UID: \"90945752-c3c4-4aaf-9680-4c5fe89ab1de\") " pod="openstack/placement-db-create-9z2kq" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.305399 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90945752-c3c4-4aaf-9680-4c5fe89ab1de-operator-scripts\") pod \"placement-db-create-9z2kq\" (UID: \"90945752-c3c4-4aaf-9680-4c5fe89ab1de\") " pod="openstack/placement-db-create-9z2kq" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.305429 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71d5f892-67e3-4854-bee6-1586d712fe8e-operator-scripts\") pod \"placement-f0fd-account-create-update-kmmtd\" (UID: \"71d5f892-67e3-4854-bee6-1586d712fe8e\") " pod="openstack/placement-f0fd-account-create-update-kmmtd" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.306892 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90945752-c3c4-4aaf-9680-4c5fe89ab1de-operator-scripts\") pod \"placement-db-create-9z2kq\" (UID: \"90945752-c3c4-4aaf-9680-4c5fe89ab1de\") " pod="openstack/placement-db-create-9z2kq" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.330173 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cnwk\" (UniqueName: \"kubernetes.io/projected/90945752-c3c4-4aaf-9680-4c5fe89ab1de-kube-api-access-8cnwk\") pod \"placement-db-create-9z2kq\" (UID: \"90945752-c3c4-4aaf-9680-4c5fe89ab1de\") " pod="openstack/placement-db-create-9z2kq" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.390191 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-9z2kq" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.408150 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71d5f892-67e3-4854-bee6-1586d712fe8e-operator-scripts\") pod \"placement-f0fd-account-create-update-kmmtd\" (UID: \"71d5f892-67e3-4854-bee6-1586d712fe8e\") " pod="openstack/placement-f0fd-account-create-update-kmmtd" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.408288 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drv8h\" (UniqueName: \"kubernetes.io/projected/71d5f892-67e3-4854-bee6-1586d712fe8e-kube-api-access-drv8h\") pod \"placement-f0fd-account-create-update-kmmtd\" (UID: \"71d5f892-67e3-4854-bee6-1586d712fe8e\") " pod="openstack/placement-f0fd-account-create-update-kmmtd" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.409234 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71d5f892-67e3-4854-bee6-1586d712fe8e-operator-scripts\") pod \"placement-f0fd-account-create-update-kmmtd\" (UID: \"71d5f892-67e3-4854-bee6-1586d712fe8e\") " pod="openstack/placement-f0fd-account-create-update-kmmtd" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.437958 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drv8h\" (UniqueName: \"kubernetes.io/projected/71d5f892-67e3-4854-bee6-1586d712fe8e-kube-api-access-drv8h\") pod \"placement-f0fd-account-create-update-kmmtd\" (UID: \"71d5f892-67e3-4854-bee6-1586d712fe8e\") " pod="openstack/placement-f0fd-account-create-update-kmmtd" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.503354 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f0fd-account-create-update-kmmtd" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.519866 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:41 crc kubenswrapper[4904]: E0214 11:25:41.520071 4904 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 14 11:25:41 crc kubenswrapper[4904]: E0214 11:25:41.520086 4904 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 14 11:25:41 crc kubenswrapper[4904]: E0214 11:25:41.520127 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift podName:0594337d-464f-489f-9761-3d3d2e143af8 nodeName:}" failed. No retries permitted until 2026-02-14 11:25:49.520113417 +0000 UTC m=+940.332878078 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift") pod "swift-storage-0" (UID: "0594337d-464f-489f-9761-3d3d2e143af8") : configmap "swift-ring-files" not found Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.699226 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-s9htl" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.705275 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8v2xr"] Feb 14 11:25:41 crc kubenswrapper[4904]: E0214 11:25:41.735637 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f52c7efb-196a-4ced-aaba-a380e99098ff" containerName="mariadb-account-create-update" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.735676 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="f52c7efb-196a-4ced-aaba-a380e99098ff" containerName="mariadb-account-create-update" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.736052 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="f52c7efb-196a-4ced-aaba-a380e99098ff" containerName="mariadb-account-create-update" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.770229 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8v2xr"] Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.770345 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8v2xr" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.791216 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-9cs22"] Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.834979 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f52c7efb-196a-4ced-aaba-a380e99098ff-operator-scripts\") pod \"f52c7efb-196a-4ced-aaba-a380e99098ff\" (UID: \"f52c7efb-196a-4ced-aaba-a380e99098ff\") " Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.835030 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsvgc\" (UniqueName: \"kubernetes.io/projected/f52c7efb-196a-4ced-aaba-a380e99098ff-kube-api-access-nsvgc\") pod \"f52c7efb-196a-4ced-aaba-a380e99098ff\" (UID: \"f52c7efb-196a-4ced-aaba-a380e99098ff\") " Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.836099 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f52c7efb-196a-4ced-aaba-a380e99098ff-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f52c7efb-196a-4ced-aaba-a380e99098ff" (UID: "f52c7efb-196a-4ced-aaba-a380e99098ff"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.865551 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f52c7efb-196a-4ced-aaba-a380e99098ff-kube-api-access-nsvgc" (OuterVolumeSpecName: "kube-api-access-nsvgc") pod "f52c7efb-196a-4ced-aaba-a380e99098ff" (UID: "f52c7efb-196a-4ced-aaba-a380e99098ff"). InnerVolumeSpecName "kube-api-access-nsvgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.887904 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-836b-account-create-update-kcjv5"] Feb 14 11:25:41 crc kubenswrapper[4904]: W0214 11:25:41.931078 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod405e6001_9b9c_46bd_8d84_d5005de91285.slice/crio-239eb47f4d73bd06fd98895cccf29d89ba01480cb407baff3277c8e1c9a27a28 WatchSource:0}: Error finding container 239eb47f4d73bd06fd98895cccf29d89ba01480cb407baff3277c8e1c9a27a28: Status 404 returned error can't find the container with id 239eb47f4d73bd06fd98895cccf29d89ba01480cb407baff3277c8e1c9a27a28 Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.936573 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw6jj\" (UniqueName: \"kubernetes.io/projected/a9fd9edc-0058-417c-baa6-9fbff5035af5-kube-api-access-qw6jj\") pod \"redhat-operators-8v2xr\" (UID: \"a9fd9edc-0058-417c-baa6-9fbff5035af5\") " pod="openshift-marketplace/redhat-operators-8v2xr" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.936639 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9fd9edc-0058-417c-baa6-9fbff5035af5-catalog-content\") pod \"redhat-operators-8v2xr\" (UID: \"a9fd9edc-0058-417c-baa6-9fbff5035af5\") " pod="openshift-marketplace/redhat-operators-8v2xr" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.936662 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9fd9edc-0058-417c-baa6-9fbff5035af5-utilities\") pod \"redhat-operators-8v2xr\" (UID: \"a9fd9edc-0058-417c-baa6-9fbff5035af5\") " pod="openshift-marketplace/redhat-operators-8v2xr" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.936763 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f52c7efb-196a-4ced-aaba-a380e99098ff-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:41 crc kubenswrapper[4904]: I0214 11:25:41.936789 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsvgc\" (UniqueName: \"kubernetes.io/projected/f52c7efb-196a-4ced-aaba-a380e99098ff-kube-api-access-nsvgc\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.026823 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-836b-account-create-update-kcjv5" event={"ID":"405e6001-9b9c-46bd-8d84-d5005de91285","Type":"ContainerStarted","Data":"239eb47f4d73bd06fd98895cccf29d89ba01480cb407baff3277c8e1c9a27a28"} Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.031482 4904 generic.go:334] "Generic (PLEG): container finished" podID="cd9c7534-8ff2-4898-ab20-660140c54857" containerID="fccd7d513406f1a26ed650c2f745b826f5bf5bdbb5a92907e09cddee575b8f43" exitCode=0 Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.031938 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4dkbs" event={"ID":"cd9c7534-8ff2-4898-ab20-660140c54857","Type":"ContainerDied","Data":"fccd7d513406f1a26ed650c2f745b826f5bf5bdbb5a92907e09cddee575b8f43"} Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.040977 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw6jj\" (UniqueName: \"kubernetes.io/projected/a9fd9edc-0058-417c-baa6-9fbff5035af5-kube-api-access-qw6jj\") pod \"redhat-operators-8v2xr\" (UID: \"a9fd9edc-0058-417c-baa6-9fbff5035af5\") " pod="openshift-marketplace/redhat-operators-8v2xr" Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.041037 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9fd9edc-0058-417c-baa6-9fbff5035af5-catalog-content\") pod \"redhat-operators-8v2xr\" (UID: \"a9fd9edc-0058-417c-baa6-9fbff5035af5\") " pod="openshift-marketplace/redhat-operators-8v2xr" Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.041062 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9fd9edc-0058-417c-baa6-9fbff5035af5-utilities\") pod \"redhat-operators-8v2xr\" (UID: \"a9fd9edc-0058-417c-baa6-9fbff5035af5\") " pod="openshift-marketplace/redhat-operators-8v2xr" Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.041536 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9fd9edc-0058-417c-baa6-9fbff5035af5-utilities\") pod \"redhat-operators-8v2xr\" (UID: \"a9fd9edc-0058-417c-baa6-9fbff5035af5\") " pod="openshift-marketplace/redhat-operators-8v2xr" Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.041943 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9fd9edc-0058-417c-baa6-9fbff5035af5-catalog-content\") pod \"redhat-operators-8v2xr\" (UID: \"a9fd9edc-0058-417c-baa6-9fbff5035af5\") " pod="openshift-marketplace/redhat-operators-8v2xr" Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.058733 4904 generic.go:334] "Generic (PLEG): container finished" podID="12e4e520-e219-4768-b5db-5ccadfee8beb" containerID="a283d1c0c7fcfd0e80f44f71d01133f59c16226f7b3998b7e82e7675ce19bb8f" exitCode=0 Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.058792 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-981b-account-create-update-rt8h8" event={"ID":"12e4e520-e219-4768-b5db-5ccadfee8beb","Type":"ContainerDied","Data":"a283d1c0c7fcfd0e80f44f71d01133f59c16226f7b3998b7e82e7675ce19bb8f"} Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.058817 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-981b-account-create-update-rt8h8" event={"ID":"12e4e520-e219-4768-b5db-5ccadfee8beb","Type":"ContainerStarted","Data":"f59aba82a53abbb2350b0478a1798271f5c014ea4a97f6dc0e9a2e9306e2c501"} Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.068323 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw6jj\" (UniqueName: \"kubernetes.io/projected/a9fd9edc-0058-417c-baa6-9fbff5035af5-kube-api-access-qw6jj\") pod \"redhat-operators-8v2xr\" (UID: \"a9fd9edc-0058-417c-baa6-9fbff5035af5\") " pod="openshift-marketplace/redhat-operators-8v2xr" Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.069417 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9cs22" event={"ID":"8649450d-1384-4d27-a4bb-501ebee69f01","Type":"ContainerStarted","Data":"e822c5c11bfc10877807ec6ddefb9563660e7887676548c7b077541be82eb6e2"} Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.080086 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-s9htl" event={"ID":"f52c7efb-196a-4ced-aaba-a380e99098ff","Type":"ContainerDied","Data":"a9ef4420644c30eb6ddd887e6c73b6f1495b8ab307cd9f58d4e55d79360f622f"} Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.080127 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9ef4420644c30eb6ddd887e6c73b6f1495b8ab307cd9f58d4e55d79360f622f" Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.080186 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-s9htl" Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.110618 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-9cs22" podStartSLOduration=2.110573206 podStartE2EDuration="2.110573206s" podCreationTimestamp="2026-02-14 11:25:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:25:42.092273715 +0000 UTC m=+932.905038376" watchObservedRunningTime="2026-02-14 11:25:42.110573206 +0000 UTC m=+932.923337867" Feb 14 11:25:42 crc kubenswrapper[4904]: E0214 11:25:42.115758 4904 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12e4e520_e219_4768_b5db_5ccadfee8beb.slice/crio-conmon-a283d1c0c7fcfd0e80f44f71d01133f59c16226f7b3998b7e82e7675ce19bb8f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf52c7efb_196a_4ced_aaba_a380e99098ff.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12e4e520_e219_4768_b5db_5ccadfee8beb.slice/crio-a283d1c0c7fcfd0e80f44f71d01133f59c16226f7b3998b7e82e7675ce19bb8f.scope\": RecentStats: unable to find data in memory cache]" Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.123487 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8v2xr" Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.191823 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-9z2kq"] Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.350704 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f0fd-account-create-update-kmmtd"] Feb 14 11:25:42 crc kubenswrapper[4904]: W0214 11:25:42.361846 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71d5f892_67e3_4854_bee6_1586d712fe8e.slice/crio-f0ecb417fd43da531da2e980ed8fed64b9bfe529050315c3026507ac19073550 WatchSource:0}: Error finding container f0ecb417fd43da531da2e980ed8fed64b9bfe529050315c3026507ac19073550: Status 404 returned error can't find the container with id f0ecb417fd43da531da2e980ed8fed64b9bfe529050315c3026507ac19073550 Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.740052 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8v2xr"] Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.756018 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.827092 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-2db8z"] Feb 14 11:25:42 crc kubenswrapper[4904]: I0214 11:25:42.827350 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" podUID="b6914d96-01d9-45d6-933a-b3b8875349df" containerName="dnsmasq-dns" containerID="cri-o://9db01e99a8ed41c20fe7746a677b1d37e6c41f5906ede9ab77f87d7f9b1b3a92" gracePeriod=10 Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.092816 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f0fd-account-create-update-kmmtd" event={"ID":"71d5f892-67e3-4854-bee6-1586d712fe8e","Type":"ContainerStarted","Data":"21efbd0cae01bcce998ef4d3fd9ca96dc2a196aa780cf0507087acc47a3ad4b1"} Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.092870 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f0fd-account-create-update-kmmtd" event={"ID":"71d5f892-67e3-4854-bee6-1586d712fe8e","Type":"ContainerStarted","Data":"f0ecb417fd43da531da2e980ed8fed64b9bfe529050315c3026507ac19073550"} Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.103181 4904 generic.go:334] "Generic (PLEG): container finished" podID="7448d80b-d340-46a2-88a6-221db1f93f11" containerID="4e2f01e425ebbd9b9326491b905f29fd60a6633fb15da5c6854ae3eebc906136" exitCode=0 Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.103235 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7448d80b-d340-46a2-88a6-221db1f93f11","Type":"ContainerDied","Data":"4e2f01e425ebbd9b9326491b905f29fd60a6633fb15da5c6854ae3eebc906136"} Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.107374 4904 generic.go:334] "Generic (PLEG): container finished" podID="a9fd9edc-0058-417c-baa6-9fbff5035af5" containerID="0d7a1d3ca53e5ca1de252c08d0b8846188358ec24880dd6a1b0426d8befd8763" exitCode=0 Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.107451 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8v2xr" event={"ID":"a9fd9edc-0058-417c-baa6-9fbff5035af5","Type":"ContainerDied","Data":"0d7a1d3ca53e5ca1de252c08d0b8846188358ec24880dd6a1b0426d8befd8763"} Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.107491 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8v2xr" event={"ID":"a9fd9edc-0058-417c-baa6-9fbff5035af5","Type":"ContainerStarted","Data":"ebd6623dce31d58bfde727984093e1af8c772c41118a2c8aa7c4d42ef1062443"} Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.113746 4904 generic.go:334] "Generic (PLEG): container finished" podID="8649450d-1384-4d27-a4bb-501ebee69f01" containerID="9216cc8fb1514cda0b04dcef5ed09bc4f644ac9951297556f53bceb35fa299f1" exitCode=0 Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.113817 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9cs22" event={"ID":"8649450d-1384-4d27-a4bb-501ebee69f01","Type":"ContainerDied","Data":"9216cc8fb1514cda0b04dcef5ed09bc4f644ac9951297556f53bceb35fa299f1"} Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.119114 4904 generic.go:334] "Generic (PLEG): container finished" podID="b6914d96-01d9-45d6-933a-b3b8875349df" containerID="9db01e99a8ed41c20fe7746a677b1d37e6c41f5906ede9ab77f87d7f9b1b3a92" exitCode=0 Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.119170 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" event={"ID":"b6914d96-01d9-45d6-933a-b3b8875349df","Type":"ContainerDied","Data":"9db01e99a8ed41c20fe7746a677b1d37e6c41f5906ede9ab77f87d7f9b1b3a92"} Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.120569 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-f0fd-account-create-update-kmmtd" podStartSLOduration=2.120558938 podStartE2EDuration="2.120558938s" podCreationTimestamp="2026-02-14 11:25:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:25:43.118753069 +0000 UTC m=+933.931517730" watchObservedRunningTime="2026-02-14 11:25:43.120558938 +0000 UTC m=+933.933323599" Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.132708 4904 generic.go:334] "Generic (PLEG): container finished" podID="405e6001-9b9c-46bd-8d84-d5005de91285" containerID="2ffffe14db50a91755915b873223a8c2218e7fb45e12f8371cf4ed51af92b2d7" exitCode=0 Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.132757 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-836b-account-create-update-kcjv5" event={"ID":"405e6001-9b9c-46bd-8d84-d5005de91285","Type":"ContainerDied","Data":"2ffffe14db50a91755915b873223a8c2218e7fb45e12f8371cf4ed51af92b2d7"} Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.139018 4904 generic.go:334] "Generic (PLEG): container finished" podID="90945752-c3c4-4aaf-9680-4c5fe89ab1de" containerID="ef630af5206762bfa2b64e289409530532689ee5af2db0e5b75aabe5596cef65" exitCode=0 Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.139134 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-9z2kq" event={"ID":"90945752-c3c4-4aaf-9680-4c5fe89ab1de","Type":"ContainerDied","Data":"ef630af5206762bfa2b64e289409530532689ee5af2db0e5b75aabe5596cef65"} Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.139157 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-9z2kq" event={"ID":"90945752-c3c4-4aaf-9680-4c5fe89ab1de","Type":"ContainerStarted","Data":"aeda67f6f6f1c471b660fffbd2af39618378bb0402168ee84a5998dbe3cd9257"} Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.329317 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-s9htl"] Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.341748 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-s9htl"] Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.526159 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.685496 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-ovsdbserver-nb\") pod \"b6914d96-01d9-45d6-933a-b3b8875349df\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.685959 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvrpk\" (UniqueName: \"kubernetes.io/projected/b6914d96-01d9-45d6-933a-b3b8875349df-kube-api-access-mvrpk\") pod \"b6914d96-01d9-45d6-933a-b3b8875349df\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.685993 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-ovsdbserver-sb\") pod \"b6914d96-01d9-45d6-933a-b3b8875349df\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.686034 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-dns-svc\") pod \"b6914d96-01d9-45d6-933a-b3b8875349df\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.686158 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-config\") pod \"b6914d96-01d9-45d6-933a-b3b8875349df\" (UID: \"b6914d96-01d9-45d6-933a-b3b8875349df\") " Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.720687 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6914d96-01d9-45d6-933a-b3b8875349df-kube-api-access-mvrpk" (OuterVolumeSpecName: "kube-api-access-mvrpk") pod "b6914d96-01d9-45d6-933a-b3b8875349df" (UID: "b6914d96-01d9-45d6-933a-b3b8875349df"). InnerVolumeSpecName "kube-api-access-mvrpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.777907 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-config" (OuterVolumeSpecName: "config") pod "b6914d96-01d9-45d6-933a-b3b8875349df" (UID: "b6914d96-01d9-45d6-933a-b3b8875349df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.788434 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvrpk\" (UniqueName: \"kubernetes.io/projected/b6914d96-01d9-45d6-933a-b3b8875349df-kube-api-access-mvrpk\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.788480 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.825198 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b6914d96-01d9-45d6-933a-b3b8875349df" (UID: "b6914d96-01d9-45d6-933a-b3b8875349df"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.831471 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b6914d96-01d9-45d6-933a-b3b8875349df" (UID: "b6914d96-01d9-45d6-933a-b3b8875349df"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.846807 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f52c7efb-196a-4ced-aaba-a380e99098ff" path="/var/lib/kubelet/pods/f52c7efb-196a-4ced-aaba-a380e99098ff/volumes" Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.876654 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b6914d96-01d9-45d6-933a-b3b8875349df" (UID: "b6914d96-01d9-45d6-933a-b3b8875349df"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.877421 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4dkbs" Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.889638 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.889669 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.889678 4904 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6914d96-01d9-45d6-933a-b3b8875349df-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.890970 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-981b-account-create-update-rt8h8" Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.990612 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqx9c\" (UniqueName: \"kubernetes.io/projected/12e4e520-e219-4768-b5db-5ccadfee8beb-kube-api-access-lqx9c\") pod \"12e4e520-e219-4768-b5db-5ccadfee8beb\" (UID: \"12e4e520-e219-4768-b5db-5ccadfee8beb\") " Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.991481 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12e4e520-e219-4768-b5db-5ccadfee8beb-operator-scripts\") pod \"12e4e520-e219-4768-b5db-5ccadfee8beb\" (UID: \"12e4e520-e219-4768-b5db-5ccadfee8beb\") " Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.991608 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd9c7534-8ff2-4898-ab20-660140c54857-operator-scripts\") pod \"cd9c7534-8ff2-4898-ab20-660140c54857\" (UID: \"cd9c7534-8ff2-4898-ab20-660140c54857\") " Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.991819 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-882qx\" (UniqueName: \"kubernetes.io/projected/cd9c7534-8ff2-4898-ab20-660140c54857-kube-api-access-882qx\") pod \"cd9c7534-8ff2-4898-ab20-660140c54857\" (UID: \"cd9c7534-8ff2-4898-ab20-660140c54857\") " Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.993180 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd9c7534-8ff2-4898-ab20-660140c54857-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cd9c7534-8ff2-4898-ab20-660140c54857" (UID: "cd9c7534-8ff2-4898-ab20-660140c54857"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.993262 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12e4e520-e219-4768-b5db-5ccadfee8beb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "12e4e520-e219-4768-b5db-5ccadfee8beb" (UID: "12e4e520-e219-4768-b5db-5ccadfee8beb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:43 crc kubenswrapper[4904]: I0214 11:25:43.998907 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd9c7534-8ff2-4898-ab20-660140c54857-kube-api-access-882qx" (OuterVolumeSpecName: "kube-api-access-882qx") pod "cd9c7534-8ff2-4898-ab20-660140c54857" (UID: "cd9c7534-8ff2-4898-ab20-660140c54857"). InnerVolumeSpecName "kube-api-access-882qx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.005322 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12e4e520-e219-4768-b5db-5ccadfee8beb-kube-api-access-lqx9c" (OuterVolumeSpecName: "kube-api-access-lqx9c") pod "12e4e520-e219-4768-b5db-5ccadfee8beb" (UID: "12e4e520-e219-4768-b5db-5ccadfee8beb"). InnerVolumeSpecName "kube-api-access-lqx9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.093649 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqx9c\" (UniqueName: \"kubernetes.io/projected/12e4e520-e219-4768-b5db-5ccadfee8beb-kube-api-access-lqx9c\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.094728 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12e4e520-e219-4768-b5db-5ccadfee8beb-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.094821 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd9c7534-8ff2-4898-ab20-660140c54857-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.094928 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-882qx\" (UniqueName: \"kubernetes.io/projected/cd9c7534-8ff2-4898-ab20-660140c54857-kube-api-access-882qx\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.147057 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8v2xr" event={"ID":"a9fd9edc-0058-417c-baa6-9fbff5035af5","Type":"ContainerStarted","Data":"2b926ba49a0893e1bd8e39df58f1784fd89df5b8d794db995e453a4fdc763e98"} Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.149640 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" event={"ID":"b6914d96-01d9-45d6-933a-b3b8875349df","Type":"ContainerDied","Data":"03e9ed03161d94a2f55f24484963ec30545014ca134be8835627d0dd67221e7f"} Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.149748 4904 scope.go:117] "RemoveContainer" containerID="9db01e99a8ed41c20fe7746a677b1d37e6c41f5906ede9ab77f87d7f9b1b3a92" Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.149920 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-2db8z" Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.156778 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4dkbs" event={"ID":"cd9c7534-8ff2-4898-ab20-660140c54857","Type":"ContainerDied","Data":"29c5cc8b5d2ba8b51f42500299abce21e4c09055c12b6fc8c66889323b9922f6"} Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.156880 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29c5cc8b5d2ba8b51f42500299abce21e4c09055c12b6fc8c66889323b9922f6" Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.156942 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4dkbs" Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.161178 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-981b-account-create-update-rt8h8" event={"ID":"12e4e520-e219-4768-b5db-5ccadfee8beb","Type":"ContainerDied","Data":"f59aba82a53abbb2350b0478a1798271f5c014ea4a97f6dc0e9a2e9306e2c501"} Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.161213 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f59aba82a53abbb2350b0478a1798271f5c014ea4a97f6dc0e9a2e9306e2c501" Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.161275 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-981b-account-create-update-rt8h8" Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.163075 4904 generic.go:334] "Generic (PLEG): container finished" podID="71d5f892-67e3-4854-bee6-1586d712fe8e" containerID="21efbd0cae01bcce998ef4d3fd9ca96dc2a196aa780cf0507087acc47a3ad4b1" exitCode=0 Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.163185 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f0fd-account-create-update-kmmtd" event={"ID":"71d5f892-67e3-4854-bee6-1586d712fe8e","Type":"ContainerDied","Data":"21efbd0cae01bcce998ef4d3fd9ca96dc2a196aa780cf0507087acc47a3ad4b1"} Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.164583 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7448d80b-d340-46a2-88a6-221db1f93f11","Type":"ContainerStarted","Data":"b2d172a2f150af5def63664c78711f02a0a8bdfe5561905a918dde89ad2f8c0d"} Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.166180 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.174972 4904 scope.go:117] "RemoveContainer" containerID="0bd481385927297abfca39eec4b813ec7ff29a4916d09495dfc078175e33205e" Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.483258 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=44.793227303 podStartE2EDuration="59.483241861s" podCreationTimestamp="2026-02-14 11:24:45 +0000 UTC" firstStartedPulling="2026-02-14 11:24:53.796651649 +0000 UTC m=+884.609416310" lastFinishedPulling="2026-02-14 11:25:08.486666207 +0000 UTC m=+899.299430868" observedRunningTime="2026-02-14 11:25:44.286217749 +0000 UTC m=+935.098982410" watchObservedRunningTime="2026-02-14 11:25:44.483241861 +0000 UTC m=+935.296006522" Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.488667 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-2db8z"] Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.506765 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-2db8z"] Feb 14 11:25:44 crc kubenswrapper[4904]: I0214 11:25:44.890183 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-836b-account-create-update-kcjv5" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.022093 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/405e6001-9b9c-46bd-8d84-d5005de91285-operator-scripts\") pod \"405e6001-9b9c-46bd-8d84-d5005de91285\" (UID: \"405e6001-9b9c-46bd-8d84-d5005de91285\") " Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.022435 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm489\" (UniqueName: \"kubernetes.io/projected/405e6001-9b9c-46bd-8d84-d5005de91285-kube-api-access-lm489\") pod \"405e6001-9b9c-46bd-8d84-d5005de91285\" (UID: \"405e6001-9b9c-46bd-8d84-d5005de91285\") " Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.031113 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/405e6001-9b9c-46bd-8d84-d5005de91285-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "405e6001-9b9c-46bd-8d84-d5005de91285" (UID: "405e6001-9b9c-46bd-8d84-d5005de91285"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.053850 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/405e6001-9b9c-46bd-8d84-d5005de91285-kube-api-access-lm489" (OuterVolumeSpecName: "kube-api-access-lm489") pod "405e6001-9b9c-46bd-8d84-d5005de91285" (UID: "405e6001-9b9c-46bd-8d84-d5005de91285"). InnerVolumeSpecName "kube-api-access-lm489". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.076611 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9cs22" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.124449 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/405e6001-9b9c-46bd-8d84-d5005de91285-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.124486 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm489\" (UniqueName: \"kubernetes.io/projected/405e6001-9b9c-46bd-8d84-d5005de91285-kube-api-access-lm489\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.188735 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9cs22" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.188741 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9cs22" event={"ID":"8649450d-1384-4d27-a4bb-501ebee69f01","Type":"ContainerDied","Data":"e822c5c11bfc10877807ec6ddefb9563660e7887676548c7b077541be82eb6e2"} Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.188767 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e822c5c11bfc10877807ec6ddefb9563660e7887676548c7b077541be82eb6e2" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.199827 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-836b-account-create-update-kcjv5" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.201719 4904 generic.go:334] "Generic (PLEG): container finished" podID="2b06dd61-da9f-455f-b813-53300e4fbc10" containerID="67f4cf2e5ed352d8d50d8a7b06dbd4929f867edd15e88f56f90d69d36544af7b" exitCode=0 Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.202878 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-836b-account-create-update-kcjv5" event={"ID":"405e6001-9b9c-46bd-8d84-d5005de91285","Type":"ContainerDied","Data":"239eb47f4d73bd06fd98895cccf29d89ba01480cb407baff3277c8e1c9a27a28"} Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.202919 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="239eb47f4d73bd06fd98895cccf29d89ba01480cb407baff3277c8e1c9a27a28" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.202937 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2b06dd61-da9f-455f-b813-53300e4fbc10","Type":"ContainerDied","Data":"67f4cf2e5ed352d8d50d8a7b06dbd4929f867edd15e88f56f90d69d36544af7b"} Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.223341 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-9z2kq" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.225390 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtp6b\" (UniqueName: \"kubernetes.io/projected/8649450d-1384-4d27-a4bb-501ebee69f01-kube-api-access-xtp6b\") pod \"8649450d-1384-4d27-a4bb-501ebee69f01\" (UID: \"8649450d-1384-4d27-a4bb-501ebee69f01\") " Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.225612 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8649450d-1384-4d27-a4bb-501ebee69f01-operator-scripts\") pod \"8649450d-1384-4d27-a4bb-501ebee69f01\" (UID: \"8649450d-1384-4d27-a4bb-501ebee69f01\") " Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.227238 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8649450d-1384-4d27-a4bb-501ebee69f01-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8649450d-1384-4d27-a4bb-501ebee69f01" (UID: "8649450d-1384-4d27-a4bb-501ebee69f01"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.251295 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8649450d-1384-4d27-a4bb-501ebee69f01-kube-api-access-xtp6b" (OuterVolumeSpecName: "kube-api-access-xtp6b") pod "8649450d-1384-4d27-a4bb-501ebee69f01" (UID: "8649450d-1384-4d27-a4bb-501ebee69f01"). InnerVolumeSpecName "kube-api-access-xtp6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.328434 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90945752-c3c4-4aaf-9680-4c5fe89ab1de-operator-scripts\") pod \"90945752-c3c4-4aaf-9680-4c5fe89ab1de\" (UID: \"90945752-c3c4-4aaf-9680-4c5fe89ab1de\") " Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.328556 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cnwk\" (UniqueName: \"kubernetes.io/projected/90945752-c3c4-4aaf-9680-4c5fe89ab1de-kube-api-access-8cnwk\") pod \"90945752-c3c4-4aaf-9680-4c5fe89ab1de\" (UID: \"90945752-c3c4-4aaf-9680-4c5fe89ab1de\") " Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.329012 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtp6b\" (UniqueName: \"kubernetes.io/projected/8649450d-1384-4d27-a4bb-501ebee69f01-kube-api-access-xtp6b\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.329042 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8649450d-1384-4d27-a4bb-501ebee69f01-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.334927 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90945752-c3c4-4aaf-9680-4c5fe89ab1de-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "90945752-c3c4-4aaf-9680-4c5fe89ab1de" (UID: "90945752-c3c4-4aaf-9680-4c5fe89ab1de"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.343086 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90945752-c3c4-4aaf-9680-4c5fe89ab1de-kube-api-access-8cnwk" (OuterVolumeSpecName: "kube-api-access-8cnwk") pod "90945752-c3c4-4aaf-9680-4c5fe89ab1de" (UID: "90945752-c3c4-4aaf-9680-4c5fe89ab1de"). InnerVolumeSpecName "kube-api-access-8cnwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.430929 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90945752-c3c4-4aaf-9680-4c5fe89ab1de-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.430957 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cnwk\" (UniqueName: \"kubernetes.io/projected/90945752-c3c4-4aaf-9680-4c5fe89ab1de-kube-api-access-8cnwk\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.652801 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f0fd-account-create-update-kmmtd" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.660469 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-j4tgk"] Feb 14 11:25:45 crc kubenswrapper[4904]: E0214 11:25:45.660768 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71d5f892-67e3-4854-bee6-1586d712fe8e" containerName="mariadb-account-create-update" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.660783 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="71d5f892-67e3-4854-bee6-1586d712fe8e" containerName="mariadb-account-create-update" Feb 14 11:25:45 crc kubenswrapper[4904]: E0214 11:25:45.660795 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6914d96-01d9-45d6-933a-b3b8875349df" containerName="init" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.660801 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6914d96-01d9-45d6-933a-b3b8875349df" containerName="init" Feb 14 11:25:45 crc kubenswrapper[4904]: E0214 11:25:45.660816 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90945752-c3c4-4aaf-9680-4c5fe89ab1de" containerName="mariadb-database-create" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.660822 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="90945752-c3c4-4aaf-9680-4c5fe89ab1de" containerName="mariadb-database-create" Feb 14 11:25:45 crc kubenswrapper[4904]: E0214 11:25:45.660829 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd9c7534-8ff2-4898-ab20-660140c54857" containerName="mariadb-database-create" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.662103 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd9c7534-8ff2-4898-ab20-660140c54857" containerName="mariadb-database-create" Feb 14 11:25:45 crc kubenswrapper[4904]: E0214 11:25:45.662118 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6914d96-01d9-45d6-933a-b3b8875349df" containerName="dnsmasq-dns" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.662123 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6914d96-01d9-45d6-933a-b3b8875349df" containerName="dnsmasq-dns" Feb 14 11:25:45 crc kubenswrapper[4904]: E0214 11:25:45.662142 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e4e520-e219-4768-b5db-5ccadfee8beb" containerName="mariadb-account-create-update" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.662148 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e4e520-e219-4768-b5db-5ccadfee8beb" containerName="mariadb-account-create-update" Feb 14 11:25:45 crc kubenswrapper[4904]: E0214 11:25:45.662160 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="405e6001-9b9c-46bd-8d84-d5005de91285" containerName="mariadb-account-create-update" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.662165 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="405e6001-9b9c-46bd-8d84-d5005de91285" containerName="mariadb-account-create-update" Feb 14 11:25:45 crc kubenswrapper[4904]: E0214 11:25:45.662173 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8649450d-1384-4d27-a4bb-501ebee69f01" containerName="mariadb-database-create" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.662178 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="8649450d-1384-4d27-a4bb-501ebee69f01" containerName="mariadb-database-create" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.662334 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="405e6001-9b9c-46bd-8d84-d5005de91285" containerName="mariadb-account-create-update" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.662345 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6914d96-01d9-45d6-933a-b3b8875349df" containerName="dnsmasq-dns" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.662355 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd9c7534-8ff2-4898-ab20-660140c54857" containerName="mariadb-database-create" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.662366 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="71d5f892-67e3-4854-bee6-1586d712fe8e" containerName="mariadb-account-create-update" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.662378 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="90945752-c3c4-4aaf-9680-4c5fe89ab1de" containerName="mariadb-database-create" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.662385 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="8649450d-1384-4d27-a4bb-501ebee69f01" containerName="mariadb-database-create" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.662393 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="12e4e520-e219-4768-b5db-5ccadfee8beb" containerName="mariadb-account-create-update" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.662891 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-j4tgk" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.665921 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.666023 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qlfk8" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.674341 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-j4tgk"] Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.745631 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drv8h\" (UniqueName: \"kubernetes.io/projected/71d5f892-67e3-4854-bee6-1586d712fe8e-kube-api-access-drv8h\") pod \"71d5f892-67e3-4854-bee6-1586d712fe8e\" (UID: \"71d5f892-67e3-4854-bee6-1586d712fe8e\") " Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.745862 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71d5f892-67e3-4854-bee6-1586d712fe8e-operator-scripts\") pod \"71d5f892-67e3-4854-bee6-1586d712fe8e\" (UID: \"71d5f892-67e3-4854-bee6-1586d712fe8e\") " Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.746126 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-combined-ca-bundle\") pod \"glance-db-sync-j4tgk\" (UID: \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\") " pod="openstack/glance-db-sync-j4tgk" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.746172 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-config-data\") pod \"glance-db-sync-j4tgk\" (UID: \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\") " pod="openstack/glance-db-sync-j4tgk" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.746347 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-db-sync-config-data\") pod \"glance-db-sync-j4tgk\" (UID: \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\") " pod="openstack/glance-db-sync-j4tgk" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.746386 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdgrf\" (UniqueName: \"kubernetes.io/projected/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-kube-api-access-tdgrf\") pod \"glance-db-sync-j4tgk\" (UID: \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\") " pod="openstack/glance-db-sync-j4tgk" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.746606 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71d5f892-67e3-4854-bee6-1586d712fe8e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "71d5f892-67e3-4854-bee6-1586d712fe8e" (UID: "71d5f892-67e3-4854-bee6-1586d712fe8e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.750167 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71d5f892-67e3-4854-bee6-1586d712fe8e-kube-api-access-drv8h" (OuterVolumeSpecName: "kube-api-access-drv8h") pod "71d5f892-67e3-4854-bee6-1586d712fe8e" (UID: "71d5f892-67e3-4854-bee6-1586d712fe8e"). InnerVolumeSpecName "kube-api-access-drv8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.844897 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6914d96-01d9-45d6-933a-b3b8875349df" path="/var/lib/kubelet/pods/b6914d96-01d9-45d6-933a-b3b8875349df/volumes" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.847762 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-db-sync-config-data\") pod \"glance-db-sync-j4tgk\" (UID: \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\") " pod="openstack/glance-db-sync-j4tgk" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.847804 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdgrf\" (UniqueName: \"kubernetes.io/projected/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-kube-api-access-tdgrf\") pod \"glance-db-sync-j4tgk\" (UID: \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\") " pod="openstack/glance-db-sync-j4tgk" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.847854 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-combined-ca-bundle\") pod \"glance-db-sync-j4tgk\" (UID: \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\") " pod="openstack/glance-db-sync-j4tgk" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.847880 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-config-data\") pod \"glance-db-sync-j4tgk\" (UID: \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\") " pod="openstack/glance-db-sync-j4tgk" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.847939 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71d5f892-67e3-4854-bee6-1586d712fe8e-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.847950 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drv8h\" (UniqueName: \"kubernetes.io/projected/71d5f892-67e3-4854-bee6-1586d712fe8e-kube-api-access-drv8h\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.851261 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-config-data\") pod \"glance-db-sync-j4tgk\" (UID: \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\") " pod="openstack/glance-db-sync-j4tgk" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.853179 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-db-sync-config-data\") pod \"glance-db-sync-j4tgk\" (UID: \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\") " pod="openstack/glance-db-sync-j4tgk" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.856293 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-combined-ca-bundle\") pod \"glance-db-sync-j4tgk\" (UID: \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\") " pod="openstack/glance-db-sync-j4tgk" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.877500 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdgrf\" (UniqueName: \"kubernetes.io/projected/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-kube-api-access-tdgrf\") pod \"glance-db-sync-j4tgk\" (UID: \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\") " pod="openstack/glance-db-sync-j4tgk" Feb 14 11:25:45 crc kubenswrapper[4904]: I0214 11:25:45.981572 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-j4tgk" Feb 14 11:25:46 crc kubenswrapper[4904]: I0214 11:25:46.225062 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2b06dd61-da9f-455f-b813-53300e4fbc10","Type":"ContainerStarted","Data":"10cf7b5b8da00578e784761802e52e4c79c9f67fb379dd1ff190c0d554e6dff3"} Feb 14 11:25:46 crc kubenswrapper[4904]: I0214 11:25:46.225716 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 14 11:25:46 crc kubenswrapper[4904]: I0214 11:25:46.228778 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-9z2kq" Feb 14 11:25:46 crc kubenswrapper[4904]: I0214 11:25:46.228856 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-9z2kq" event={"ID":"90945752-c3c4-4aaf-9680-4c5fe89ab1de","Type":"ContainerDied","Data":"aeda67f6f6f1c471b660fffbd2af39618378bb0402168ee84a5998dbe3cd9257"} Feb 14 11:25:46 crc kubenswrapper[4904]: I0214 11:25:46.228892 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aeda67f6f6f1c471b660fffbd2af39618378bb0402168ee84a5998dbe3cd9257" Feb 14 11:25:46 crc kubenswrapper[4904]: I0214 11:25:46.230696 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f0fd-account-create-update-kmmtd" event={"ID":"71d5f892-67e3-4854-bee6-1586d712fe8e","Type":"ContainerDied","Data":"f0ecb417fd43da531da2e980ed8fed64b9bfe529050315c3026507ac19073550"} Feb 14 11:25:46 crc kubenswrapper[4904]: I0214 11:25:46.230715 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f0fd-account-create-update-kmmtd" Feb 14 11:25:46 crc kubenswrapper[4904]: I0214 11:25:46.230727 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0ecb417fd43da531da2e980ed8fed64b9bfe529050315c3026507ac19073550" Feb 14 11:25:46 crc kubenswrapper[4904]: I0214 11:25:46.296105 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=40.257407242 podStartE2EDuration="1m1.296088175s" podCreationTimestamp="2026-02-14 11:24:45 +0000 UTC" firstStartedPulling="2026-02-14 11:24:47.448057027 +0000 UTC m=+878.260821678" lastFinishedPulling="2026-02-14 11:25:08.48673795 +0000 UTC m=+899.299502611" observedRunningTime="2026-02-14 11:25:46.287898371 +0000 UTC m=+937.100663052" watchObservedRunningTime="2026-02-14 11:25:46.296088175 +0000 UTC m=+937.108852836" Feb 14 11:25:46 crc kubenswrapper[4904]: I0214 11:25:46.383082 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:25:46 crc kubenswrapper[4904]: I0214 11:25:46.383135 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:25:46 crc kubenswrapper[4904]: I0214 11:25:46.460066 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-j4tgk"] Feb 14 11:25:46 crc kubenswrapper[4904]: W0214 11:25:46.468514 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod042a40bf_ba7e_4f1a_afb9_a4f82a3470e1.slice/crio-17fc667c28b358c070356f3e1967adedff6b9e1d79be4711e87f5ac328d099b4 WatchSource:0}: Error finding container 17fc667c28b358c070356f3e1967adedff6b9e1d79be4711e87f5ac328d099b4: Status 404 returned error can't find the container with id 17fc667c28b358c070356f3e1967adedff6b9e1d79be4711e87f5ac328d099b4 Feb 14 11:25:47 crc kubenswrapper[4904]: I0214 11:25:47.127539 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-ts9fx"] Feb 14 11:25:47 crc kubenswrapper[4904]: I0214 11:25:47.128739 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-ts9fx" Feb 14 11:25:47 crc kubenswrapper[4904]: I0214 11:25:47.133409 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 14 11:25:47 crc kubenswrapper[4904]: I0214 11:25:47.138068 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-ts9fx"] Feb 14 11:25:47 crc kubenswrapper[4904]: I0214 11:25:47.170908 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bc80ca6-8741-476d-b7a5-99eef2609a6d-operator-scripts\") pod \"root-account-create-update-ts9fx\" (UID: \"2bc80ca6-8741-476d-b7a5-99eef2609a6d\") " pod="openstack/root-account-create-update-ts9fx" Feb 14 11:25:47 crc kubenswrapper[4904]: I0214 11:25:47.170992 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfm68\" (UniqueName: \"kubernetes.io/projected/2bc80ca6-8741-476d-b7a5-99eef2609a6d-kube-api-access-bfm68\") pod \"root-account-create-update-ts9fx\" (UID: \"2bc80ca6-8741-476d-b7a5-99eef2609a6d\") " pod="openstack/root-account-create-update-ts9fx" Feb 14 11:25:47 crc kubenswrapper[4904]: I0214 11:25:47.238343 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-j4tgk" event={"ID":"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1","Type":"ContainerStarted","Data":"17fc667c28b358c070356f3e1967adedff6b9e1d79be4711e87f5ac328d099b4"} Feb 14 11:25:47 crc kubenswrapper[4904]: I0214 11:25:47.240579 4904 generic.go:334] "Generic (PLEG): container finished" podID="a9fd9edc-0058-417c-baa6-9fbff5035af5" containerID="2b926ba49a0893e1bd8e39df58f1784fd89df5b8d794db995e453a4fdc763e98" exitCode=0 Feb 14 11:25:47 crc kubenswrapper[4904]: I0214 11:25:47.241388 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8v2xr" event={"ID":"a9fd9edc-0058-417c-baa6-9fbff5035af5","Type":"ContainerDied","Data":"2b926ba49a0893e1bd8e39df58f1784fd89df5b8d794db995e453a4fdc763e98"} Feb 14 11:25:47 crc kubenswrapper[4904]: I0214 11:25:47.272004 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfm68\" (UniqueName: \"kubernetes.io/projected/2bc80ca6-8741-476d-b7a5-99eef2609a6d-kube-api-access-bfm68\") pod \"root-account-create-update-ts9fx\" (UID: \"2bc80ca6-8741-476d-b7a5-99eef2609a6d\") " pod="openstack/root-account-create-update-ts9fx" Feb 14 11:25:47 crc kubenswrapper[4904]: I0214 11:25:47.272110 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bc80ca6-8741-476d-b7a5-99eef2609a6d-operator-scripts\") pod \"root-account-create-update-ts9fx\" (UID: \"2bc80ca6-8741-476d-b7a5-99eef2609a6d\") " pod="openstack/root-account-create-update-ts9fx" Feb 14 11:25:47 crc kubenswrapper[4904]: I0214 11:25:47.272774 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bc80ca6-8741-476d-b7a5-99eef2609a6d-operator-scripts\") pod \"root-account-create-update-ts9fx\" (UID: \"2bc80ca6-8741-476d-b7a5-99eef2609a6d\") " pod="openstack/root-account-create-update-ts9fx" Feb 14 11:25:47 crc kubenswrapper[4904]: I0214 11:25:47.299091 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfm68\" (UniqueName: \"kubernetes.io/projected/2bc80ca6-8741-476d-b7a5-99eef2609a6d-kube-api-access-bfm68\") pod \"root-account-create-update-ts9fx\" (UID: \"2bc80ca6-8741-476d-b7a5-99eef2609a6d\") " pod="openstack/root-account-create-update-ts9fx" Feb 14 11:25:47 crc kubenswrapper[4904]: I0214 11:25:47.443252 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-ts9fx" Feb 14 11:25:48 crc kubenswrapper[4904]: I0214 11:25:48.001962 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-ts9fx"] Feb 14 11:25:48 crc kubenswrapper[4904]: I0214 11:25:48.004961 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Feb 14 11:25:48 crc kubenswrapper[4904]: I0214 11:25:48.250911 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8v2xr" event={"ID":"a9fd9edc-0058-417c-baa6-9fbff5035af5","Type":"ContainerStarted","Data":"522c8e0477ec7ed39163c6dc7ca5c2542ed209a4caa41b362756bc571a8ad301"} Feb 14 11:25:48 crc kubenswrapper[4904]: I0214 11:25:48.253719 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-ts9fx" event={"ID":"2bc80ca6-8741-476d-b7a5-99eef2609a6d","Type":"ContainerStarted","Data":"9f345d8012a3bbd34736da7a1d8efa129d14b0c2092b8099c1cb6addc6a8c5ca"} Feb 14 11:25:48 crc kubenswrapper[4904]: I0214 11:25:48.253747 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-ts9fx" event={"ID":"2bc80ca6-8741-476d-b7a5-99eef2609a6d","Type":"ContainerStarted","Data":"731b32bd75cffbe93683622480d9df753a9f8843bd1f1a8217b5413960ffb07a"} Feb 14 11:25:48 crc kubenswrapper[4904]: I0214 11:25:48.272943 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8v2xr" podStartSLOduration=2.649894301 podStartE2EDuration="7.272927636s" podCreationTimestamp="2026-02-14 11:25:41 +0000 UTC" firstStartedPulling="2026-02-14 11:25:43.109950088 +0000 UTC m=+933.922714749" lastFinishedPulling="2026-02-14 11:25:47.732983423 +0000 UTC m=+938.545748084" observedRunningTime="2026-02-14 11:25:48.269274097 +0000 UTC m=+939.082038758" watchObservedRunningTime="2026-02-14 11:25:48.272927636 +0000 UTC m=+939.085692297" Feb 14 11:25:48 crc kubenswrapper[4904]: I0214 11:25:48.294091 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-ts9fx" podStartSLOduration=1.294073416 podStartE2EDuration="1.294073416s" podCreationTimestamp="2026-02-14 11:25:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:25:48.287294801 +0000 UTC m=+939.100059462" watchObservedRunningTime="2026-02-14 11:25:48.294073416 +0000 UTC m=+939.106838077" Feb 14 11:25:49 crc kubenswrapper[4904]: I0214 11:25:49.270911 4904 generic.go:334] "Generic (PLEG): container finished" podID="2bc80ca6-8741-476d-b7a5-99eef2609a6d" containerID="9f345d8012a3bbd34736da7a1d8efa129d14b0c2092b8099c1cb6addc6a8c5ca" exitCode=0 Feb 14 11:25:49 crc kubenswrapper[4904]: I0214 11:25:49.270950 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-ts9fx" event={"ID":"2bc80ca6-8741-476d-b7a5-99eef2609a6d","Type":"ContainerDied","Data":"9f345d8012a3bbd34736da7a1d8efa129d14b0c2092b8099c1cb6addc6a8c5ca"} Feb 14 11:25:49 crc kubenswrapper[4904]: I0214 11:25:49.622903 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:25:49 crc kubenswrapper[4904]: E0214 11:25:49.623114 4904 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 14 11:25:49 crc kubenswrapper[4904]: E0214 11:25:49.623143 4904 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 14 11:25:49 crc kubenswrapper[4904]: E0214 11:25:49.623197 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift podName:0594337d-464f-489f-9761-3d3d2e143af8 nodeName:}" failed. No retries permitted until 2026-02-14 11:26:05.623181479 +0000 UTC m=+956.435946140 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift") pod "swift-storage-0" (UID: "0594337d-464f-489f-9761-3d3d2e143af8") : configmap "swift-ring-files" not found Feb 14 11:25:50 crc kubenswrapper[4904]: I0214 11:25:50.311503 4904 generic.go:334] "Generic (PLEG): container finished" podID="e2a79513-6beb-47b8-96e5-b5424056661e" containerID="a99499067fec11baa46e626f0ebc36bea23368382a810a4827c0d937c2719e48" exitCode=0 Feb 14 11:25:50 crc kubenswrapper[4904]: I0214 11:25:50.311960 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tw4ls" event={"ID":"e2a79513-6beb-47b8-96e5-b5424056661e","Type":"ContainerDied","Data":"a99499067fec11baa46e626f0ebc36bea23368382a810a4827c0d937c2719e48"} Feb 14 11:25:50 crc kubenswrapper[4904]: I0214 11:25:50.642786 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-ts9fx" Feb 14 11:25:50 crc kubenswrapper[4904]: I0214 11:25:50.755207 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfm68\" (UniqueName: \"kubernetes.io/projected/2bc80ca6-8741-476d-b7a5-99eef2609a6d-kube-api-access-bfm68\") pod \"2bc80ca6-8741-476d-b7a5-99eef2609a6d\" (UID: \"2bc80ca6-8741-476d-b7a5-99eef2609a6d\") " Feb 14 11:25:50 crc kubenswrapper[4904]: I0214 11:25:50.755351 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bc80ca6-8741-476d-b7a5-99eef2609a6d-operator-scripts\") pod \"2bc80ca6-8741-476d-b7a5-99eef2609a6d\" (UID: \"2bc80ca6-8741-476d-b7a5-99eef2609a6d\") " Feb 14 11:25:50 crc kubenswrapper[4904]: I0214 11:25:50.756503 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bc80ca6-8741-476d-b7a5-99eef2609a6d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2bc80ca6-8741-476d-b7a5-99eef2609a6d" (UID: "2bc80ca6-8741-476d-b7a5-99eef2609a6d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:50 crc kubenswrapper[4904]: I0214 11:25:50.766579 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bc80ca6-8741-476d-b7a5-99eef2609a6d-kube-api-access-bfm68" (OuterVolumeSpecName: "kube-api-access-bfm68") pod "2bc80ca6-8741-476d-b7a5-99eef2609a6d" (UID: "2bc80ca6-8741-476d-b7a5-99eef2609a6d"). InnerVolumeSpecName "kube-api-access-bfm68". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:25:50 crc kubenswrapper[4904]: I0214 11:25:50.857131 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfm68\" (UniqueName: \"kubernetes.io/projected/2bc80ca6-8741-476d-b7a5-99eef2609a6d-kube-api-access-bfm68\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:50 crc kubenswrapper[4904]: I0214 11:25:50.857158 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bc80ca6-8741-476d-b7a5-99eef2609a6d-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:51 crc kubenswrapper[4904]: I0214 11:25:51.323715 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-ts9fx" Feb 14 11:25:51 crc kubenswrapper[4904]: I0214 11:25:51.332565 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-ts9fx" event={"ID":"2bc80ca6-8741-476d-b7a5-99eef2609a6d","Type":"ContainerDied","Data":"731b32bd75cffbe93683622480d9df753a9f8843bd1f1a8217b5413960ffb07a"} Feb 14 11:25:51 crc kubenswrapper[4904]: I0214 11:25:51.332603 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="731b32bd75cffbe93683622480d9df753a9f8843bd1f1a8217b5413960ffb07a" Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.101998 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.125077 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8v2xr" Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.125116 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8v2xr" Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.280091 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2a79513-6beb-47b8-96e5-b5424056661e-scripts\") pod \"e2a79513-6beb-47b8-96e5-b5424056661e\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.280127 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2s9dz\" (UniqueName: \"kubernetes.io/projected/e2a79513-6beb-47b8-96e5-b5424056661e-kube-api-access-2s9dz\") pod \"e2a79513-6beb-47b8-96e5-b5424056661e\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.280160 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2a79513-6beb-47b8-96e5-b5424056661e-dispersionconf\") pod \"e2a79513-6beb-47b8-96e5-b5424056661e\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.280204 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2a79513-6beb-47b8-96e5-b5424056661e-swiftconf\") pod \"e2a79513-6beb-47b8-96e5-b5424056661e\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.280271 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2a79513-6beb-47b8-96e5-b5424056661e-etc-swift\") pod \"e2a79513-6beb-47b8-96e5-b5424056661e\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.280353 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2a79513-6beb-47b8-96e5-b5424056661e-combined-ca-bundle\") pod \"e2a79513-6beb-47b8-96e5-b5424056661e\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.280400 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2a79513-6beb-47b8-96e5-b5424056661e-ring-data-devices\") pod \"e2a79513-6beb-47b8-96e5-b5424056661e\" (UID: \"e2a79513-6beb-47b8-96e5-b5424056661e\") " Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.286752 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2a79513-6beb-47b8-96e5-b5424056661e-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e2a79513-6beb-47b8-96e5-b5424056661e" (UID: "e2a79513-6beb-47b8-96e5-b5424056661e"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.287305 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2a79513-6beb-47b8-96e5-b5424056661e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e2a79513-6beb-47b8-96e5-b5424056661e" (UID: "e2a79513-6beb-47b8-96e5-b5424056661e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.288041 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2a79513-6beb-47b8-96e5-b5424056661e-kube-api-access-2s9dz" (OuterVolumeSpecName: "kube-api-access-2s9dz") pod "e2a79513-6beb-47b8-96e5-b5424056661e" (UID: "e2a79513-6beb-47b8-96e5-b5424056661e"). InnerVolumeSpecName "kube-api-access-2s9dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.309748 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2a79513-6beb-47b8-96e5-b5424056661e-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e2a79513-6beb-47b8-96e5-b5424056661e" (UID: "e2a79513-6beb-47b8-96e5-b5424056661e"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.318596 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2a79513-6beb-47b8-96e5-b5424056661e-scripts" (OuterVolumeSpecName: "scripts") pod "e2a79513-6beb-47b8-96e5-b5424056661e" (UID: "e2a79513-6beb-47b8-96e5-b5424056661e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.322002 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2a79513-6beb-47b8-96e5-b5424056661e-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e2a79513-6beb-47b8-96e5-b5424056661e" (UID: "e2a79513-6beb-47b8-96e5-b5424056661e"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.328318 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2a79513-6beb-47b8-96e5-b5424056661e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2a79513-6beb-47b8-96e5-b5424056661e" (UID: "e2a79513-6beb-47b8-96e5-b5424056661e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.337101 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tw4ls" event={"ID":"e2a79513-6beb-47b8-96e5-b5424056661e","Type":"ContainerDied","Data":"bdb42885c87db0ed09b0d05217ad3125bcdcddb2071442bb14d180181d97d752"} Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.337142 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdb42885c87db0ed09b0d05217ad3125bcdcddb2071442bb14d180181d97d752" Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.337210 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tw4ls" Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.383104 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2a79513-6beb-47b8-96e5-b5424056661e-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.383140 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2s9dz\" (UniqueName: \"kubernetes.io/projected/e2a79513-6beb-47b8-96e5-b5424056661e-kube-api-access-2s9dz\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.383152 4904 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2a79513-6beb-47b8-96e5-b5424056661e-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.383172 4904 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2a79513-6beb-47b8-96e5-b5424056661e-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.383181 4904 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2a79513-6beb-47b8-96e5-b5424056661e-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.383189 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2a79513-6beb-47b8-96e5-b5424056661e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:52 crc kubenswrapper[4904]: I0214 11:25:52.383198 4904 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2a79513-6beb-47b8-96e5-b5424056661e-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 14 11:25:53 crc kubenswrapper[4904]: I0214 11:25:53.178177 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8v2xr" podUID="a9fd9edc-0058-417c-baa6-9fbff5035af5" containerName="registry-server" probeResult="failure" output=< Feb 14 11:25:53 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 11:25:53 crc kubenswrapper[4904]: > Feb 14 11:25:53 crc kubenswrapper[4904]: I0214 11:25:53.293583 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-ts9fx"] Feb 14 11:25:53 crc kubenswrapper[4904]: I0214 11:25:53.300364 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-ts9fx"] Feb 14 11:25:53 crc kubenswrapper[4904]: I0214 11:25:53.845733 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bc80ca6-8741-476d-b7a5-99eef2609a6d" path="/var/lib/kubelet/pods/2bc80ca6-8741-476d-b7a5-99eef2609a6d/volumes" Feb 14 11:25:54 crc kubenswrapper[4904]: I0214 11:25:54.921932 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-zj52b" podUID="05a401e2-caea-48a9-bb53-6a42b59dad57" containerName="ovn-controller" probeResult="failure" output=< Feb 14 11:25:54 crc kubenswrapper[4904]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 14 11:25:54 crc kubenswrapper[4904]: > Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.018279 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.020376 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-9jk9m" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.291039 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zj52b-config-g2d7g"] Feb 14 11:25:55 crc kubenswrapper[4904]: E0214 11:25:55.291364 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bc80ca6-8741-476d-b7a5-99eef2609a6d" containerName="mariadb-account-create-update" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.291383 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bc80ca6-8741-476d-b7a5-99eef2609a6d" containerName="mariadb-account-create-update" Feb 14 11:25:55 crc kubenswrapper[4904]: E0214 11:25:55.291393 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2a79513-6beb-47b8-96e5-b5424056661e" containerName="swift-ring-rebalance" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.291401 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2a79513-6beb-47b8-96e5-b5424056661e" containerName="swift-ring-rebalance" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.291579 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bc80ca6-8741-476d-b7a5-99eef2609a6d" containerName="mariadb-account-create-update" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.291596 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2a79513-6beb-47b8-96e5-b5424056661e" containerName="swift-ring-rebalance" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.292130 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.294937 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.314151 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zj52b-config-g2d7g"] Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.484615 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt7mr\" (UniqueName: \"kubernetes.io/projected/0239db54-31c1-4a95-b280-e814bbd07d5c-kube-api-access-mt7mr\") pod \"ovn-controller-zj52b-config-g2d7g\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.484677 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0239db54-31c1-4a95-b280-e814bbd07d5c-var-log-ovn\") pod \"ovn-controller-zj52b-config-g2d7g\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.484736 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0239db54-31c1-4a95-b280-e814bbd07d5c-additional-scripts\") pod \"ovn-controller-zj52b-config-g2d7g\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.484783 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0239db54-31c1-4a95-b280-e814bbd07d5c-var-run\") pod \"ovn-controller-zj52b-config-g2d7g\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.484867 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0239db54-31c1-4a95-b280-e814bbd07d5c-var-run-ovn\") pod \"ovn-controller-zj52b-config-g2d7g\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.484897 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0239db54-31c1-4a95-b280-e814bbd07d5c-scripts\") pod \"ovn-controller-zj52b-config-g2d7g\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.585866 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt7mr\" (UniqueName: \"kubernetes.io/projected/0239db54-31c1-4a95-b280-e814bbd07d5c-kube-api-access-mt7mr\") pod \"ovn-controller-zj52b-config-g2d7g\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.585920 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0239db54-31c1-4a95-b280-e814bbd07d5c-var-log-ovn\") pod \"ovn-controller-zj52b-config-g2d7g\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.585972 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0239db54-31c1-4a95-b280-e814bbd07d5c-additional-scripts\") pod \"ovn-controller-zj52b-config-g2d7g\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.586035 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0239db54-31c1-4a95-b280-e814bbd07d5c-var-run\") pod \"ovn-controller-zj52b-config-g2d7g\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.586085 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0239db54-31c1-4a95-b280-e814bbd07d5c-var-run-ovn\") pod \"ovn-controller-zj52b-config-g2d7g\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.586108 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0239db54-31c1-4a95-b280-e814bbd07d5c-scripts\") pod \"ovn-controller-zj52b-config-g2d7g\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.588035 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0239db54-31c1-4a95-b280-e814bbd07d5c-scripts\") pod \"ovn-controller-zj52b-config-g2d7g\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.589251 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0239db54-31c1-4a95-b280-e814bbd07d5c-var-run\") pod \"ovn-controller-zj52b-config-g2d7g\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.589866 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0239db54-31c1-4a95-b280-e814bbd07d5c-additional-scripts\") pod \"ovn-controller-zj52b-config-g2d7g\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.589931 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0239db54-31c1-4a95-b280-e814bbd07d5c-var-run-ovn\") pod \"ovn-controller-zj52b-config-g2d7g\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.590196 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0239db54-31c1-4a95-b280-e814bbd07d5c-var-log-ovn\") pod \"ovn-controller-zj52b-config-g2d7g\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.631041 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt7mr\" (UniqueName: \"kubernetes.io/projected/0239db54-31c1-4a95-b280-e814bbd07d5c-kube-api-access-mt7mr\") pod \"ovn-controller-zj52b-config-g2d7g\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:55 crc kubenswrapper[4904]: I0214 11:25:55.907806 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:25:56 crc kubenswrapper[4904]: I0214 11:25:56.316964 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zj52b-config-g2d7g"] Feb 14 11:25:56 crc kubenswrapper[4904]: I0214 11:25:56.410054 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zj52b-config-g2d7g" event={"ID":"0239db54-31c1-4a95-b280-e814bbd07d5c","Type":"ContainerStarted","Data":"97992087e2d3edbc2af0b905876f3e3ab699be730e017fe0e5f60ae7f8419291"} Feb 14 11:25:56 crc kubenswrapper[4904]: I0214 11:25:56.731761 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="2b06dd61-da9f-455f-b813-53300e4fbc10" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.94:5671: connect: connection refused" Feb 14 11:25:57 crc kubenswrapper[4904]: I0214 11:25:57.135470 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="7448d80b-d340-46a2-88a6-221db1f93f11" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.95:5671: connect: connection refused" Feb 14 11:25:57 crc kubenswrapper[4904]: I0214 11:25:57.419123 4904 generic.go:334] "Generic (PLEG): container finished" podID="0239db54-31c1-4a95-b280-e814bbd07d5c" containerID="a2103801ef86e1b867adb135f0a9513174a6c5cc3638429966de7427d29c4333" exitCode=0 Feb 14 11:25:57 crc kubenswrapper[4904]: I0214 11:25:57.419171 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zj52b-config-g2d7g" event={"ID":"0239db54-31c1-4a95-b280-e814bbd07d5c","Type":"ContainerDied","Data":"a2103801ef86e1b867adb135f0a9513174a6c5cc3638429966de7427d29c4333"} Feb 14 11:25:58 crc kubenswrapper[4904]: I0214 11:25:58.310139 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-c4294"] Feb 14 11:25:58 crc kubenswrapper[4904]: I0214 11:25:58.311342 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-c4294" Feb 14 11:25:58 crc kubenswrapper[4904]: I0214 11:25:58.317997 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 14 11:25:58 crc kubenswrapper[4904]: I0214 11:25:58.322572 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-c4294"] Feb 14 11:25:58 crc kubenswrapper[4904]: I0214 11:25:58.453010 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73518149-3e47-413e-8bf1-c29279ee89b7-operator-scripts\") pod \"root-account-create-update-c4294\" (UID: \"73518149-3e47-413e-8bf1-c29279ee89b7\") " pod="openstack/root-account-create-update-c4294" Feb 14 11:25:58 crc kubenswrapper[4904]: I0214 11:25:58.453361 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjfmc\" (UniqueName: \"kubernetes.io/projected/73518149-3e47-413e-8bf1-c29279ee89b7-kube-api-access-wjfmc\") pod \"root-account-create-update-c4294\" (UID: \"73518149-3e47-413e-8bf1-c29279ee89b7\") " pod="openstack/root-account-create-update-c4294" Feb 14 11:25:58 crc kubenswrapper[4904]: I0214 11:25:58.555394 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73518149-3e47-413e-8bf1-c29279ee89b7-operator-scripts\") pod \"root-account-create-update-c4294\" (UID: \"73518149-3e47-413e-8bf1-c29279ee89b7\") " pod="openstack/root-account-create-update-c4294" Feb 14 11:25:58 crc kubenswrapper[4904]: I0214 11:25:58.555459 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjfmc\" (UniqueName: \"kubernetes.io/projected/73518149-3e47-413e-8bf1-c29279ee89b7-kube-api-access-wjfmc\") pod \"root-account-create-update-c4294\" (UID: \"73518149-3e47-413e-8bf1-c29279ee89b7\") " pod="openstack/root-account-create-update-c4294" Feb 14 11:25:58 crc kubenswrapper[4904]: I0214 11:25:58.556755 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73518149-3e47-413e-8bf1-c29279ee89b7-operator-scripts\") pod \"root-account-create-update-c4294\" (UID: \"73518149-3e47-413e-8bf1-c29279ee89b7\") " pod="openstack/root-account-create-update-c4294" Feb 14 11:25:58 crc kubenswrapper[4904]: I0214 11:25:58.584465 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjfmc\" (UniqueName: \"kubernetes.io/projected/73518149-3e47-413e-8bf1-c29279ee89b7-kube-api-access-wjfmc\") pod \"root-account-create-update-c4294\" (UID: \"73518149-3e47-413e-8bf1-c29279ee89b7\") " pod="openstack/root-account-create-update-c4294" Feb 14 11:25:58 crc kubenswrapper[4904]: I0214 11:25:58.627005 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-c4294" Feb 14 11:25:59 crc kubenswrapper[4904]: I0214 11:25:59.895464 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-zj52b" Feb 14 11:26:03 crc kubenswrapper[4904]: I0214 11:26:03.174617 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8v2xr" podUID="a9fd9edc-0058-417c-baa6-9fbff5035af5" containerName="registry-server" probeResult="failure" output=< Feb 14 11:26:03 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 11:26:03 crc kubenswrapper[4904]: > Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.702725 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.708528 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0594337d-464f-489f-9761-3d3d2e143af8-etc-swift\") pod \"swift-storage-0\" (UID: \"0594337d-464f-489f-9761-3d3d2e143af8\") " pod="openstack/swift-storage-0" Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.750524 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.761244 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.906637 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0239db54-31c1-4a95-b280-e814bbd07d5c-additional-scripts\") pod \"0239db54-31c1-4a95-b280-e814bbd07d5c\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.908151 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0239db54-31c1-4a95-b280-e814bbd07d5c-scripts\") pod \"0239db54-31c1-4a95-b280-e814bbd07d5c\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.908246 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mt7mr\" (UniqueName: \"kubernetes.io/projected/0239db54-31c1-4a95-b280-e814bbd07d5c-kube-api-access-mt7mr\") pod \"0239db54-31c1-4a95-b280-e814bbd07d5c\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.908294 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0239db54-31c1-4a95-b280-e814bbd07d5c-var-run-ovn\") pod \"0239db54-31c1-4a95-b280-e814bbd07d5c\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.908301 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0239db54-31c1-4a95-b280-e814bbd07d5c-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "0239db54-31c1-4a95-b280-e814bbd07d5c" (UID: "0239db54-31c1-4a95-b280-e814bbd07d5c"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.908364 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0239db54-31c1-4a95-b280-e814bbd07d5c-var-log-ovn\") pod \"0239db54-31c1-4a95-b280-e814bbd07d5c\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.908380 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0239db54-31c1-4a95-b280-e814bbd07d5c-var-run\") pod \"0239db54-31c1-4a95-b280-e814bbd07d5c\" (UID: \"0239db54-31c1-4a95-b280-e814bbd07d5c\") " Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.908486 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0239db54-31c1-4a95-b280-e814bbd07d5c-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "0239db54-31c1-4a95-b280-e814bbd07d5c" (UID: "0239db54-31c1-4a95-b280-e814bbd07d5c"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.908578 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0239db54-31c1-4a95-b280-e814bbd07d5c-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "0239db54-31c1-4a95-b280-e814bbd07d5c" (UID: "0239db54-31c1-4a95-b280-e814bbd07d5c"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.908603 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0239db54-31c1-4a95-b280-e814bbd07d5c-var-run" (OuterVolumeSpecName: "var-run") pod "0239db54-31c1-4a95-b280-e814bbd07d5c" (UID: "0239db54-31c1-4a95-b280-e814bbd07d5c"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.908760 4904 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0239db54-31c1-4a95-b280-e814bbd07d5c-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.908771 4904 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0239db54-31c1-4a95-b280-e814bbd07d5c-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.908781 4904 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0239db54-31c1-4a95-b280-e814bbd07d5c-var-run\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.908788 4904 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0239db54-31c1-4a95-b280-e814bbd07d5c-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.909191 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0239db54-31c1-4a95-b280-e814bbd07d5c-scripts" (OuterVolumeSpecName: "scripts") pod "0239db54-31c1-4a95-b280-e814bbd07d5c" (UID: "0239db54-31c1-4a95-b280-e814bbd07d5c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:05 crc kubenswrapper[4904]: I0214 11:26:05.913497 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0239db54-31c1-4a95-b280-e814bbd07d5c-kube-api-access-mt7mr" (OuterVolumeSpecName: "kube-api-access-mt7mr") pod "0239db54-31c1-4a95-b280-e814bbd07d5c" (UID: "0239db54-31c1-4a95-b280-e814bbd07d5c"). InnerVolumeSpecName "kube-api-access-mt7mr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:06 crc kubenswrapper[4904]: I0214 11:26:06.012926 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0239db54-31c1-4a95-b280-e814bbd07d5c-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:06 crc kubenswrapper[4904]: I0214 11:26:06.012959 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mt7mr\" (UniqueName: \"kubernetes.io/projected/0239db54-31c1-4a95-b280-e814bbd07d5c-kube-api-access-mt7mr\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:06 crc kubenswrapper[4904]: I0214 11:26:06.093877 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-c4294"] Feb 14 11:26:06 crc kubenswrapper[4904]: I0214 11:26:06.394216 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 14 11:26:06 crc kubenswrapper[4904]: W0214 11:26:06.401958 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0594337d_464f_489f_9761_3d3d2e143af8.slice/crio-1dfeb896112289572c97243d3115615449f83023d2c20ec4cf717837c1c65192 WatchSource:0}: Error finding container 1dfeb896112289572c97243d3115615449f83023d2c20ec4cf717837c1c65192: Status 404 returned error can't find the container with id 1dfeb896112289572c97243d3115615449f83023d2c20ec4cf717837c1c65192 Feb 14 11:26:06 crc kubenswrapper[4904]: I0214 11:26:06.498793 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-c4294" event={"ID":"73518149-3e47-413e-8bf1-c29279ee89b7","Type":"ContainerStarted","Data":"6f5ffacb5f290b1aab1e4dbef99e1d5b80188276c4a1162896b14f1d0342ea14"} Feb 14 11:26:06 crc kubenswrapper[4904]: I0214 11:26:06.498855 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-c4294" event={"ID":"73518149-3e47-413e-8bf1-c29279ee89b7","Type":"ContainerStarted","Data":"f0226c51e932fc0cdb16c02a09f78e5bb6132919880a557af07f83784c920ebe"} Feb 14 11:26:06 crc kubenswrapper[4904]: I0214 11:26:06.502233 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-j4tgk" event={"ID":"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1","Type":"ContainerStarted","Data":"ac18286729f4904e4ad9a3194b4bf19ab0d8831192515607e99661486cdffff4"} Feb 14 11:26:06 crc kubenswrapper[4904]: I0214 11:26:06.504942 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0594337d-464f-489f-9761-3d3d2e143af8","Type":"ContainerStarted","Data":"1dfeb896112289572c97243d3115615449f83023d2c20ec4cf717837c1c65192"} Feb 14 11:26:06 crc kubenswrapper[4904]: I0214 11:26:06.506603 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zj52b-config-g2d7g" event={"ID":"0239db54-31c1-4a95-b280-e814bbd07d5c","Type":"ContainerDied","Data":"97992087e2d3edbc2af0b905876f3e3ab699be730e017fe0e5f60ae7f8419291"} Feb 14 11:26:06 crc kubenswrapper[4904]: I0214 11:26:06.506638 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97992087e2d3edbc2af0b905876f3e3ab699be730e017fe0e5f60ae7f8419291" Feb 14 11:26:06 crc kubenswrapper[4904]: I0214 11:26:06.506699 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zj52b-config-g2d7g" Feb 14 11:26:06 crc kubenswrapper[4904]: I0214 11:26:06.523226 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-c4294" podStartSLOduration=8.523210446 podStartE2EDuration="8.523210446s" podCreationTimestamp="2026-02-14 11:25:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:26:06.516155822 +0000 UTC m=+957.328920483" watchObservedRunningTime="2026-02-14 11:26:06.523210446 +0000 UTC m=+957.335975097" Feb 14 11:26:06 crc kubenswrapper[4904]: I0214 11:26:06.545513 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-j4tgk" podStartSLOduration=2.342370703 podStartE2EDuration="21.545496917s" podCreationTimestamp="2026-02-14 11:25:45 +0000 UTC" firstStartedPulling="2026-02-14 11:25:46.470810636 +0000 UTC m=+937.283575297" lastFinishedPulling="2026-02-14 11:26:05.67393685 +0000 UTC m=+956.486701511" observedRunningTime="2026-02-14 11:26:06.539491342 +0000 UTC m=+957.352256003" watchObservedRunningTime="2026-02-14 11:26:06.545496917 +0000 UTC m=+957.358261578" Feb 14 11:26:06 crc kubenswrapper[4904]: I0214 11:26:06.731064 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 14 11:26:06 crc kubenswrapper[4904]: I0214 11:26:06.936718 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-zj52b-config-g2d7g"] Feb 14 11:26:06 crc kubenswrapper[4904]: I0214 11:26:06.948859 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-zj52b-config-g2d7g"] Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.140093 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.183499 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-vfmnn"] Feb 14 11:26:07 crc kubenswrapper[4904]: E0214 11:26:07.187034 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0239db54-31c1-4a95-b280-e814bbd07d5c" containerName="ovn-config" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.187066 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="0239db54-31c1-4a95-b280-e814bbd07d5c" containerName="ovn-config" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.187296 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="0239db54-31c1-4a95-b280-e814bbd07d5c" containerName="ovn-config" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.187821 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vfmnn" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.205184 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vfmnn"] Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.350268 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzrf2\" (UniqueName: \"kubernetes.io/projected/e148cac5-499e-488a-b06b-9aedaa6f2b41-kube-api-access-fzrf2\") pod \"cinder-db-create-vfmnn\" (UID: \"e148cac5-499e-488a-b06b-9aedaa6f2b41\") " pod="openstack/cinder-db-create-vfmnn" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.350357 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e148cac5-499e-488a-b06b-9aedaa6f2b41-operator-scripts\") pod \"cinder-db-create-vfmnn\" (UID: \"e148cac5-499e-488a-b06b-9aedaa6f2b41\") " pod="openstack/cinder-db-create-vfmnn" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.451687 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzrf2\" (UniqueName: \"kubernetes.io/projected/e148cac5-499e-488a-b06b-9aedaa6f2b41-kube-api-access-fzrf2\") pod \"cinder-db-create-vfmnn\" (UID: \"e148cac5-499e-488a-b06b-9aedaa6f2b41\") " pod="openstack/cinder-db-create-vfmnn" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.451768 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e148cac5-499e-488a-b06b-9aedaa6f2b41-operator-scripts\") pod \"cinder-db-create-vfmnn\" (UID: \"e148cac5-499e-488a-b06b-9aedaa6f2b41\") " pod="openstack/cinder-db-create-vfmnn" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.452435 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e148cac5-499e-488a-b06b-9aedaa6f2b41-operator-scripts\") pod \"cinder-db-create-vfmnn\" (UID: \"e148cac5-499e-488a-b06b-9aedaa6f2b41\") " pod="openstack/cinder-db-create-vfmnn" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.469015 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-8fad-account-create-update-k8rxj"] Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.469991 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8fad-account-create-update-k8rxj" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.497100 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-8fad-account-create-update-k8rxj"] Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.499592 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.542693 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzrf2\" (UniqueName: \"kubernetes.io/projected/e148cac5-499e-488a-b06b-9aedaa6f2b41-kube-api-access-fzrf2\") pod \"cinder-db-create-vfmnn\" (UID: \"e148cac5-499e-488a-b06b-9aedaa6f2b41\") " pod="openstack/cinder-db-create-vfmnn" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.553139 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84nrn\" (UniqueName: \"kubernetes.io/projected/a628e9c6-8ff3-41bb-92ae-69d703f224da-kube-api-access-84nrn\") pod \"cinder-8fad-account-create-update-k8rxj\" (UID: \"a628e9c6-8ff3-41bb-92ae-69d703f224da\") " pod="openstack/cinder-8fad-account-create-update-k8rxj" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.553324 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a628e9c6-8ff3-41bb-92ae-69d703f224da-operator-scripts\") pod \"cinder-8fad-account-create-update-k8rxj\" (UID: \"a628e9c6-8ff3-41bb-92ae-69d703f224da\") " pod="openstack/cinder-8fad-account-create-update-k8rxj" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.558677 4904 generic.go:334] "Generic (PLEG): container finished" podID="73518149-3e47-413e-8bf1-c29279ee89b7" containerID="6f5ffacb5f290b1aab1e4dbef99e1d5b80188276c4a1162896b14f1d0342ea14" exitCode=0 Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.559040 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-c4294" event={"ID":"73518149-3e47-413e-8bf1-c29279ee89b7","Type":"ContainerDied","Data":"6f5ffacb5f290b1aab1e4dbef99e1d5b80188276c4a1162896b14f1d0342ea14"} Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.654350 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a628e9c6-8ff3-41bb-92ae-69d703f224da-operator-scripts\") pod \"cinder-8fad-account-create-update-k8rxj\" (UID: \"a628e9c6-8ff3-41bb-92ae-69d703f224da\") " pod="openstack/cinder-8fad-account-create-update-k8rxj" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.654638 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84nrn\" (UniqueName: \"kubernetes.io/projected/a628e9c6-8ff3-41bb-92ae-69d703f224da-kube-api-access-84nrn\") pod \"cinder-8fad-account-create-update-k8rxj\" (UID: \"a628e9c6-8ff3-41bb-92ae-69d703f224da\") " pod="openstack/cinder-8fad-account-create-update-k8rxj" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.656049 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a628e9c6-8ff3-41bb-92ae-69d703f224da-operator-scripts\") pod \"cinder-8fad-account-create-update-k8rxj\" (UID: \"a628e9c6-8ff3-41bb-92ae-69d703f224da\") " pod="openstack/cinder-8fad-account-create-update-k8rxj" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.688824 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84nrn\" (UniqueName: \"kubernetes.io/projected/a628e9c6-8ff3-41bb-92ae-69d703f224da-kube-api-access-84nrn\") pod \"cinder-8fad-account-create-update-k8rxj\" (UID: \"a628e9c6-8ff3-41bb-92ae-69d703f224da\") " pod="openstack/cinder-8fad-account-create-update-k8rxj" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.748792 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-7m2cz"] Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.755324 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7m2cz" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.796498 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8fad-account-create-update-k8rxj" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.799791 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-7m2cz"] Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.808449 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vfmnn" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.863550 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10c49b28-99e8-4e52-9950-bfec689cc77c-operator-scripts\") pod \"barbican-db-create-7m2cz\" (UID: \"10c49b28-99e8-4e52-9950-bfec689cc77c\") " pod="openstack/barbican-db-create-7m2cz" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.863643 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf6j6\" (UniqueName: \"kubernetes.io/projected/10c49b28-99e8-4e52-9950-bfec689cc77c-kube-api-access-rf6j6\") pod \"barbican-db-create-7m2cz\" (UID: \"10c49b28-99e8-4e52-9950-bfec689cc77c\") " pod="openstack/barbican-db-create-7m2cz" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.878927 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0239db54-31c1-4a95-b280-e814bbd07d5c" path="/var/lib/kubelet/pods/0239db54-31c1-4a95-b280-e814bbd07d5c/volumes" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.887911 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-9ldqj"] Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.889097 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9ldqj" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.906957 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9ldqj"] Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.934820 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-8b70-account-create-update-j2fhg"] Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.935861 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8b70-account-create-update-j2fhg" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.949113 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.965865 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2xb7\" (UniqueName: \"kubernetes.io/projected/4168b6f1-2a3a-410e-a0c6-42783873ab95-kube-api-access-n2xb7\") pod \"barbican-8b70-account-create-update-j2fhg\" (UID: \"4168b6f1-2a3a-410e-a0c6-42783873ab95\") " pod="openstack/barbican-8b70-account-create-update-j2fhg" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.966109 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10c49b28-99e8-4e52-9950-bfec689cc77c-operator-scripts\") pod \"barbican-db-create-7m2cz\" (UID: \"10c49b28-99e8-4e52-9950-bfec689cc77c\") " pod="openstack/barbican-db-create-7m2cz" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.966216 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9215ece4-8699-4e9a-aac3-1c1d2b6822c2-operator-scripts\") pod \"neutron-db-create-9ldqj\" (UID: \"9215ece4-8699-4e9a-aac3-1c1d2b6822c2\") " pod="openstack/neutron-db-create-9ldqj" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.966306 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgxdd\" (UniqueName: \"kubernetes.io/projected/9215ece4-8699-4e9a-aac3-1c1d2b6822c2-kube-api-access-bgxdd\") pod \"neutron-db-create-9ldqj\" (UID: \"9215ece4-8699-4e9a-aac3-1c1d2b6822c2\") " pod="openstack/neutron-db-create-9ldqj" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.966377 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf6j6\" (UniqueName: \"kubernetes.io/projected/10c49b28-99e8-4e52-9950-bfec689cc77c-kube-api-access-rf6j6\") pod \"barbican-db-create-7m2cz\" (UID: \"10c49b28-99e8-4e52-9950-bfec689cc77c\") " pod="openstack/barbican-db-create-7m2cz" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.966447 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4168b6f1-2a3a-410e-a0c6-42783873ab95-operator-scripts\") pod \"barbican-8b70-account-create-update-j2fhg\" (UID: \"4168b6f1-2a3a-410e-a0c6-42783873ab95\") " pod="openstack/barbican-8b70-account-create-update-j2fhg" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.967204 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10c49b28-99e8-4e52-9950-bfec689cc77c-operator-scripts\") pod \"barbican-db-create-7m2cz\" (UID: \"10c49b28-99e8-4e52-9950-bfec689cc77c\") " pod="openstack/barbican-db-create-7m2cz" Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.984212 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8b70-account-create-update-j2fhg"] Feb 14 11:26:07 crc kubenswrapper[4904]: I0214 11:26:07.996611 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-jc69g"] Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.002472 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jc69g" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.008766 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.011014 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.011461 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-57x7t" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.023131 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.023488 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf6j6\" (UniqueName: \"kubernetes.io/projected/10c49b28-99e8-4e52-9950-bfec689cc77c-kube-api-access-rf6j6\") pod \"barbican-db-create-7m2cz\" (UID: \"10c49b28-99e8-4e52-9950-bfec689cc77c\") " pod="openstack/barbican-db-create-7m2cz" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.047021 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jc69g"] Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.071359 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2xb7\" (UniqueName: \"kubernetes.io/projected/4168b6f1-2a3a-410e-a0c6-42783873ab95-kube-api-access-n2xb7\") pod \"barbican-8b70-account-create-update-j2fhg\" (UID: \"4168b6f1-2a3a-410e-a0c6-42783873ab95\") " pod="openstack/barbican-8b70-account-create-update-j2fhg" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.071883 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9215ece4-8699-4e9a-aac3-1c1d2b6822c2-operator-scripts\") pod \"neutron-db-create-9ldqj\" (UID: \"9215ece4-8699-4e9a-aac3-1c1d2b6822c2\") " pod="openstack/neutron-db-create-9ldqj" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.072015 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgxdd\" (UniqueName: \"kubernetes.io/projected/9215ece4-8699-4e9a-aac3-1c1d2b6822c2-kube-api-access-bgxdd\") pod \"neutron-db-create-9ldqj\" (UID: \"9215ece4-8699-4e9a-aac3-1c1d2b6822c2\") " pod="openstack/neutron-db-create-9ldqj" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.072099 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a1a642-2b0a-4a5c-825f-883de77215e2-combined-ca-bundle\") pod \"keystone-db-sync-jc69g\" (UID: \"e9a1a642-2b0a-4a5c-825f-883de77215e2\") " pod="openstack/keystone-db-sync-jc69g" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.072178 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4168b6f1-2a3a-410e-a0c6-42783873ab95-operator-scripts\") pod \"barbican-8b70-account-create-update-j2fhg\" (UID: \"4168b6f1-2a3a-410e-a0c6-42783873ab95\") " pod="openstack/barbican-8b70-account-create-update-j2fhg" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.072270 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9a1a642-2b0a-4a5c-825f-883de77215e2-config-data\") pod \"keystone-db-sync-jc69g\" (UID: \"e9a1a642-2b0a-4a5c-825f-883de77215e2\") " pod="openstack/keystone-db-sync-jc69g" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.072354 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6fzz\" (UniqueName: \"kubernetes.io/projected/e9a1a642-2b0a-4a5c-825f-883de77215e2-kube-api-access-v6fzz\") pod \"keystone-db-sync-jc69g\" (UID: \"e9a1a642-2b0a-4a5c-825f-883de77215e2\") " pod="openstack/keystone-db-sync-jc69g" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.073135 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4168b6f1-2a3a-410e-a0c6-42783873ab95-operator-scripts\") pod \"barbican-8b70-account-create-update-j2fhg\" (UID: \"4168b6f1-2a3a-410e-a0c6-42783873ab95\") " pod="openstack/barbican-8b70-account-create-update-j2fhg" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.073969 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9215ece4-8699-4e9a-aac3-1c1d2b6822c2-operator-scripts\") pod \"neutron-db-create-9ldqj\" (UID: \"9215ece4-8699-4e9a-aac3-1c1d2b6822c2\") " pod="openstack/neutron-db-create-9ldqj" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.095932 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgxdd\" (UniqueName: \"kubernetes.io/projected/9215ece4-8699-4e9a-aac3-1c1d2b6822c2-kube-api-access-bgxdd\") pod \"neutron-db-create-9ldqj\" (UID: \"9215ece4-8699-4e9a-aac3-1c1d2b6822c2\") " pod="openstack/neutron-db-create-9ldqj" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.106414 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2xb7\" (UniqueName: \"kubernetes.io/projected/4168b6f1-2a3a-410e-a0c6-42783873ab95-kube-api-access-n2xb7\") pod \"barbican-8b70-account-create-update-j2fhg\" (UID: \"4168b6f1-2a3a-410e-a0c6-42783873ab95\") " pod="openstack/barbican-8b70-account-create-update-j2fhg" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.128597 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7m2cz" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.147037 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dbdb-account-create-update-6drwj"] Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.148010 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dbdb-account-create-update-6drwj" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.152192 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.177120 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a1a642-2b0a-4a5c-825f-883de77215e2-combined-ca-bundle\") pod \"keystone-db-sync-jc69g\" (UID: \"e9a1a642-2b0a-4a5c-825f-883de77215e2\") " pod="openstack/keystone-db-sync-jc69g" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.177159 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60a86662-8426-41d7-8eab-bd0ed8ff9fda-operator-scripts\") pod \"neutron-dbdb-account-create-update-6drwj\" (UID: \"60a86662-8426-41d7-8eab-bd0ed8ff9fda\") " pod="openstack/neutron-dbdb-account-create-update-6drwj" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.177195 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv8bh\" (UniqueName: \"kubernetes.io/projected/60a86662-8426-41d7-8eab-bd0ed8ff9fda-kube-api-access-sv8bh\") pod \"neutron-dbdb-account-create-update-6drwj\" (UID: \"60a86662-8426-41d7-8eab-bd0ed8ff9fda\") " pod="openstack/neutron-dbdb-account-create-update-6drwj" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.177233 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9a1a642-2b0a-4a5c-825f-883de77215e2-config-data\") pod \"keystone-db-sync-jc69g\" (UID: \"e9a1a642-2b0a-4a5c-825f-883de77215e2\") " pod="openstack/keystone-db-sync-jc69g" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.177264 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6fzz\" (UniqueName: \"kubernetes.io/projected/e9a1a642-2b0a-4a5c-825f-883de77215e2-kube-api-access-v6fzz\") pod \"keystone-db-sync-jc69g\" (UID: \"e9a1a642-2b0a-4a5c-825f-883de77215e2\") " pod="openstack/keystone-db-sync-jc69g" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.184152 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dbdb-account-create-update-6drwj"] Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.185570 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a1a642-2b0a-4a5c-825f-883de77215e2-combined-ca-bundle\") pod \"keystone-db-sync-jc69g\" (UID: \"e9a1a642-2b0a-4a5c-825f-883de77215e2\") " pod="openstack/keystone-db-sync-jc69g" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.197520 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9a1a642-2b0a-4a5c-825f-883de77215e2-config-data\") pod \"keystone-db-sync-jc69g\" (UID: \"e9a1a642-2b0a-4a5c-825f-883de77215e2\") " pod="openstack/keystone-db-sync-jc69g" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.225902 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6fzz\" (UniqueName: \"kubernetes.io/projected/e9a1a642-2b0a-4a5c-825f-883de77215e2-kube-api-access-v6fzz\") pod \"keystone-db-sync-jc69g\" (UID: \"e9a1a642-2b0a-4a5c-825f-883de77215e2\") " pod="openstack/keystone-db-sync-jc69g" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.251781 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9ldqj" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.286104 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60a86662-8426-41d7-8eab-bd0ed8ff9fda-operator-scripts\") pod \"neutron-dbdb-account-create-update-6drwj\" (UID: \"60a86662-8426-41d7-8eab-bd0ed8ff9fda\") " pod="openstack/neutron-dbdb-account-create-update-6drwj" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.286157 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv8bh\" (UniqueName: \"kubernetes.io/projected/60a86662-8426-41d7-8eab-bd0ed8ff9fda-kube-api-access-sv8bh\") pod \"neutron-dbdb-account-create-update-6drwj\" (UID: \"60a86662-8426-41d7-8eab-bd0ed8ff9fda\") " pod="openstack/neutron-dbdb-account-create-update-6drwj" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.287110 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60a86662-8426-41d7-8eab-bd0ed8ff9fda-operator-scripts\") pod \"neutron-dbdb-account-create-update-6drwj\" (UID: \"60a86662-8426-41d7-8eab-bd0ed8ff9fda\") " pod="openstack/neutron-dbdb-account-create-update-6drwj" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.314624 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv8bh\" (UniqueName: \"kubernetes.io/projected/60a86662-8426-41d7-8eab-bd0ed8ff9fda-kube-api-access-sv8bh\") pod \"neutron-dbdb-account-create-update-6drwj\" (UID: \"60a86662-8426-41d7-8eab-bd0ed8ff9fda\") " pod="openstack/neutron-dbdb-account-create-update-6drwj" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.365444 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8b70-account-create-update-j2fhg" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.390544 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jc69g" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.498196 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dbdb-account-create-update-6drwj" Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.815379 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-8fad-account-create-update-k8rxj"] Feb 14 11:26:08 crc kubenswrapper[4904]: I0214 11:26:08.941454 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-7m2cz"] Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.116927 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9ldqj"] Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.120988 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8b70-account-create-update-j2fhg"] Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.279496 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vfmnn"] Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.301788 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jc69g"] Feb 14 11:26:09 crc kubenswrapper[4904]: W0214 11:26:09.372901 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9215ece4_8699_4e9a_aac3_1c1d2b6822c2.slice/crio-31b01af65725bed4faed9e4ae06e42f89843d120490d3790d2296c4aae0406d3 WatchSource:0}: Error finding container 31b01af65725bed4faed9e4ae06e42f89843d120490d3790d2296c4aae0406d3: Status 404 returned error can't find the container with id 31b01af65725bed4faed9e4ae06e42f89843d120490d3790d2296c4aae0406d3 Feb 14 11:26:09 crc kubenswrapper[4904]: W0214 11:26:09.373439 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda628e9c6_8ff3_41bb_92ae_69d703f224da.slice/crio-0978eb3a719b63da5040d182b949b09be2ebba442474f6fdfddcbb97d695bd81 WatchSource:0}: Error finding container 0978eb3a719b63da5040d182b949b09be2ebba442474f6fdfddcbb97d695bd81: Status 404 returned error can't find the container with id 0978eb3a719b63da5040d182b949b09be2ebba442474f6fdfddcbb97d695bd81 Feb 14 11:26:09 crc kubenswrapper[4904]: W0214 11:26:09.376768 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10c49b28_99e8_4e52_9950_bfec689cc77c.slice/crio-5d1dd0a60e559a175869ca09c521ad854e2fcf89239960d5db04cd5b9a79a383 WatchSource:0}: Error finding container 5d1dd0a60e559a175869ca09c521ad854e2fcf89239960d5db04cd5b9a79a383: Status 404 returned error can't find the container with id 5d1dd0a60e559a175869ca09c521ad854e2fcf89239960d5db04cd5b9a79a383 Feb 14 11:26:09 crc kubenswrapper[4904]: W0214 11:26:09.379448 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4168b6f1_2a3a_410e_a0c6_42783873ab95.slice/crio-e467cef0b586f2fa61b2d66de7c2304391effa8f1a3c4b326e129234633c95f8 WatchSource:0}: Error finding container e467cef0b586f2fa61b2d66de7c2304391effa8f1a3c4b326e129234633c95f8: Status 404 returned error can't find the container with id e467cef0b586f2fa61b2d66de7c2304391effa8f1a3c4b326e129234633c95f8 Feb 14 11:26:09 crc kubenswrapper[4904]: W0214 11:26:09.382464 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9a1a642_2b0a_4a5c_825f_883de77215e2.slice/crio-58d8d4fd4d8d1c3014dda653bd847d5e9088534b77ab149a1dec761647276cf2 WatchSource:0}: Error finding container 58d8d4fd4d8d1c3014dda653bd847d5e9088534b77ab149a1dec761647276cf2: Status 404 returned error can't find the container with id 58d8d4fd4d8d1c3014dda653bd847d5e9088534b77ab149a1dec761647276cf2 Feb 14 11:26:09 crc kubenswrapper[4904]: W0214 11:26:09.383246 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode148cac5_499e_488a_b06b_9aedaa6f2b41.slice/crio-ffd24fa912ddfa84b15920ac561d15d66707308590e726f075a0dd8aca4aa7bb WatchSource:0}: Error finding container ffd24fa912ddfa84b15920ac561d15d66707308590e726f075a0dd8aca4aa7bb: Status 404 returned error can't find the container with id ffd24fa912ddfa84b15920ac561d15d66707308590e726f075a0dd8aca4aa7bb Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.469765 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-c4294" Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.516095 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjfmc\" (UniqueName: \"kubernetes.io/projected/73518149-3e47-413e-8bf1-c29279ee89b7-kube-api-access-wjfmc\") pod \"73518149-3e47-413e-8bf1-c29279ee89b7\" (UID: \"73518149-3e47-413e-8bf1-c29279ee89b7\") " Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.516204 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73518149-3e47-413e-8bf1-c29279ee89b7-operator-scripts\") pod \"73518149-3e47-413e-8bf1-c29279ee89b7\" (UID: \"73518149-3e47-413e-8bf1-c29279ee89b7\") " Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.517337 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73518149-3e47-413e-8bf1-c29279ee89b7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "73518149-3e47-413e-8bf1-c29279ee89b7" (UID: "73518149-3e47-413e-8bf1-c29279ee89b7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.530128 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73518149-3e47-413e-8bf1-c29279ee89b7-kube-api-access-wjfmc" (OuterVolumeSpecName: "kube-api-access-wjfmc") pod "73518149-3e47-413e-8bf1-c29279ee89b7" (UID: "73518149-3e47-413e-8bf1-c29279ee89b7"). InnerVolumeSpecName "kube-api-access-wjfmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.610190 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8b70-account-create-update-j2fhg" event={"ID":"4168b6f1-2a3a-410e-a0c6-42783873ab95","Type":"ContainerStarted","Data":"e467cef0b586f2fa61b2d66de7c2304391effa8f1a3c4b326e129234633c95f8"} Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.610876 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jc69g" event={"ID":"e9a1a642-2b0a-4a5c-825f-883de77215e2","Type":"ContainerStarted","Data":"58d8d4fd4d8d1c3014dda653bd847d5e9088534b77ab149a1dec761647276cf2"} Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.611767 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-c4294" event={"ID":"73518149-3e47-413e-8bf1-c29279ee89b7","Type":"ContainerDied","Data":"f0226c51e932fc0cdb16c02a09f78e5bb6132919880a557af07f83784c920ebe"} Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.611791 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0226c51e932fc0cdb16c02a09f78e5bb6132919880a557af07f83784c920ebe" Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.611851 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-c4294" Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.630067 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8fad-account-create-update-k8rxj" event={"ID":"a628e9c6-8ff3-41bb-92ae-69d703f224da","Type":"ContainerStarted","Data":"0978eb3a719b63da5040d182b949b09be2ebba442474f6fdfddcbb97d695bd81"} Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.630823 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73518149-3e47-413e-8bf1-c29279ee89b7-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.630874 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjfmc\" (UniqueName: \"kubernetes.io/projected/73518149-3e47-413e-8bf1-c29279ee89b7-kube-api-access-wjfmc\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.643375 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7m2cz" event={"ID":"10c49b28-99e8-4e52-9950-bfec689cc77c","Type":"ContainerStarted","Data":"5d1dd0a60e559a175869ca09c521ad854e2fcf89239960d5db04cd5b9a79a383"} Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.653720 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9ldqj" event={"ID":"9215ece4-8699-4e9a-aac3-1c1d2b6822c2","Type":"ContainerStarted","Data":"31b01af65725bed4faed9e4ae06e42f89843d120490d3790d2296c4aae0406d3"} Feb 14 11:26:09 crc kubenswrapper[4904]: I0214 11:26:09.655304 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vfmnn" event={"ID":"e148cac5-499e-488a-b06b-9aedaa6f2b41","Type":"ContainerStarted","Data":"ffd24fa912ddfa84b15920ac561d15d66707308590e726f075a0dd8aca4aa7bb"} Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:10.066903 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dbdb-account-create-update-6drwj"] Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:10.206597 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:10.672718 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dbdb-account-create-update-6drwj" event={"ID":"60a86662-8426-41d7-8eab-bd0ed8ff9fda","Type":"ContainerStarted","Data":"ff27b00163379c4bd4576662061e384aaf59c646dc7b1d4b17d0c032ac02b086"} Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:10.676533 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8fad-account-create-update-k8rxj" event={"ID":"a628e9c6-8ff3-41bb-92ae-69d703f224da","Type":"ContainerStarted","Data":"33cbf2577b882f6c3e5c3df634cb48a4602198a44f88474756c48719932734a3"} Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:10.683515 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7m2cz" event={"ID":"10c49b28-99e8-4e52-9950-bfec689cc77c","Type":"ContainerStarted","Data":"bd7ace9778e88cca437bf1ade30faea54060444a0a0b802788627437c3702387"} Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:10.690859 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0594337d-464f-489f-9761-3d3d2e143af8","Type":"ContainerStarted","Data":"c678294f813673da1d319388e3a8d79fdeca2f659739696944f7743301bd92b8"} Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:10.715622 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9ldqj" event={"ID":"9215ece4-8699-4e9a-aac3-1c1d2b6822c2","Type":"ContainerStarted","Data":"cf77dce6159e3dfeb40683ce7ba91a167973488c00ea7c742cd4fbad2ffb935c"} Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:10.728691 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vfmnn" event={"ID":"e148cac5-499e-488a-b06b-9aedaa6f2b41","Type":"ContainerStarted","Data":"60584e082f5e434f71f5e7be8e1dec7daffc3a3759a146c30a5a4f18560e78ab"} Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:10.736141 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8b70-account-create-update-j2fhg" event={"ID":"4168b6f1-2a3a-410e-a0c6-42783873ab95","Type":"ContainerStarted","Data":"6446be203b1275df99c9c26e3aef8f97bbcab08aadbc5021232ba082c4fa7b66"} Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:10.740022 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-7m2cz" podStartSLOduration=3.739985652 podStartE2EDuration="3.739985652s" podCreationTimestamp="2026-02-14 11:26:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:26:10.72203529 +0000 UTC m=+961.534799951" watchObservedRunningTime="2026-02-14 11:26:10.739985652 +0000 UTC m=+961.552750313" Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:10.758811 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-9ldqj" podStartSLOduration=3.758796578 podStartE2EDuration="3.758796578s" podCreationTimestamp="2026-02-14 11:26:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:26:10.756377411 +0000 UTC m=+961.569142072" watchObservedRunningTime="2026-02-14 11:26:10.758796578 +0000 UTC m=+961.571561239" Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:10.788324 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-vfmnn" podStartSLOduration=3.788308307 podStartE2EDuration="3.788308307s" podCreationTimestamp="2026-02-14 11:26:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:26:10.775797944 +0000 UTC m=+961.588562605" watchObservedRunningTime="2026-02-14 11:26:10.788308307 +0000 UTC m=+961.601072968" Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:10.797820 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-8b70-account-create-update-j2fhg" podStartSLOduration=3.797807027 podStartE2EDuration="3.797807027s" podCreationTimestamp="2026-02-14 11:26:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:26:10.795534205 +0000 UTC m=+961.608298856" watchObservedRunningTime="2026-02-14 11:26:10.797807027 +0000 UTC m=+961.610571688" Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:11.752172 4904 generic.go:334] "Generic (PLEG): container finished" podID="9215ece4-8699-4e9a-aac3-1c1d2b6822c2" containerID="cf77dce6159e3dfeb40683ce7ba91a167973488c00ea7c742cd4fbad2ffb935c" exitCode=0 Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:11.752375 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9ldqj" event={"ID":"9215ece4-8699-4e9a-aac3-1c1d2b6822c2","Type":"ContainerDied","Data":"cf77dce6159e3dfeb40683ce7ba91a167973488c00ea7c742cd4fbad2ffb935c"} Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:11.760468 4904 generic.go:334] "Generic (PLEG): container finished" podID="e148cac5-499e-488a-b06b-9aedaa6f2b41" containerID="60584e082f5e434f71f5e7be8e1dec7daffc3a3759a146c30a5a4f18560e78ab" exitCode=0 Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:11.760626 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vfmnn" event={"ID":"e148cac5-499e-488a-b06b-9aedaa6f2b41","Type":"ContainerDied","Data":"60584e082f5e434f71f5e7be8e1dec7daffc3a3759a146c30a5a4f18560e78ab"} Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:11.763857 4904 generic.go:334] "Generic (PLEG): container finished" podID="4168b6f1-2a3a-410e-a0c6-42783873ab95" containerID="6446be203b1275df99c9c26e3aef8f97bbcab08aadbc5021232ba082c4fa7b66" exitCode=0 Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:11.763925 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8b70-account-create-update-j2fhg" event={"ID":"4168b6f1-2a3a-410e-a0c6-42783873ab95","Type":"ContainerDied","Data":"6446be203b1275df99c9c26e3aef8f97bbcab08aadbc5021232ba082c4fa7b66"} Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:11.771033 4904 generic.go:334] "Generic (PLEG): container finished" podID="a628e9c6-8ff3-41bb-92ae-69d703f224da" containerID="33cbf2577b882f6c3e5c3df634cb48a4602198a44f88474756c48719932734a3" exitCode=0 Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:11.771105 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8fad-account-create-update-k8rxj" event={"ID":"a628e9c6-8ff3-41bb-92ae-69d703f224da","Type":"ContainerDied","Data":"33cbf2577b882f6c3e5c3df634cb48a4602198a44f88474756c48719932734a3"} Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:11.774726 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dbdb-account-create-update-6drwj" event={"ID":"60a86662-8426-41d7-8eab-bd0ed8ff9fda","Type":"ContainerStarted","Data":"063736057a86ef2533b8d53d185c3959a9c8feed0f08ea11c82cc5ffeeb6c0e0"} Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:11.777854 4904 generic.go:334] "Generic (PLEG): container finished" podID="10c49b28-99e8-4e52-9950-bfec689cc77c" containerID="bd7ace9778e88cca437bf1ade30faea54060444a0a0b802788627437c3702387" exitCode=0 Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:11.781650 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7m2cz" event={"ID":"10c49b28-99e8-4e52-9950-bfec689cc77c","Type":"ContainerDied","Data":"bd7ace9778e88cca437bf1ade30faea54060444a0a0b802788627437c3702387"} Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:11.783939 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0594337d-464f-489f-9761-3d3d2e143af8","Type":"ContainerStarted","Data":"996d0294136cbbe8a20a4093383adfe9ecdbe5fd079c2e1f57f062da836ab368"} Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:11.783958 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0594337d-464f-489f-9761-3d3d2e143af8","Type":"ContainerStarted","Data":"23064aad486227c23e99227b14409da871212f3fa524ed63aab74eea21bc2163"} Feb 14 11:26:11 crc kubenswrapper[4904]: I0214 11:26:11.921950 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dbdb-account-create-update-6drwj" podStartSLOduration=3.921935919 podStartE2EDuration="3.921935919s" podCreationTimestamp="2026-02-14 11:26:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:26:11.889463438 +0000 UTC m=+962.702228099" watchObservedRunningTime="2026-02-14 11:26:11.921935919 +0000 UTC m=+962.734700580" Feb 14 11:26:12 crc kubenswrapper[4904]: I0214 11:26:12.332492 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8fad-account-create-update-k8rxj" Feb 14 11:26:12 crc kubenswrapper[4904]: I0214 11:26:12.466281 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84nrn\" (UniqueName: \"kubernetes.io/projected/a628e9c6-8ff3-41bb-92ae-69d703f224da-kube-api-access-84nrn\") pod \"a628e9c6-8ff3-41bb-92ae-69d703f224da\" (UID: \"a628e9c6-8ff3-41bb-92ae-69d703f224da\") " Feb 14 11:26:12 crc kubenswrapper[4904]: I0214 11:26:12.466440 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a628e9c6-8ff3-41bb-92ae-69d703f224da-operator-scripts\") pod \"a628e9c6-8ff3-41bb-92ae-69d703f224da\" (UID: \"a628e9c6-8ff3-41bb-92ae-69d703f224da\") " Feb 14 11:26:12 crc kubenswrapper[4904]: I0214 11:26:12.466924 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a628e9c6-8ff3-41bb-92ae-69d703f224da-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a628e9c6-8ff3-41bb-92ae-69d703f224da" (UID: "a628e9c6-8ff3-41bb-92ae-69d703f224da"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:12 crc kubenswrapper[4904]: I0214 11:26:12.472174 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a628e9c6-8ff3-41bb-92ae-69d703f224da-kube-api-access-84nrn" (OuterVolumeSpecName: "kube-api-access-84nrn") pod "a628e9c6-8ff3-41bb-92ae-69d703f224da" (UID: "a628e9c6-8ff3-41bb-92ae-69d703f224da"). InnerVolumeSpecName "kube-api-access-84nrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:12 crc kubenswrapper[4904]: I0214 11:26:12.568451 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a628e9c6-8ff3-41bb-92ae-69d703f224da-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:12 crc kubenswrapper[4904]: I0214 11:26:12.568829 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84nrn\" (UniqueName: \"kubernetes.io/projected/a628e9c6-8ff3-41bb-92ae-69d703f224da-kube-api-access-84nrn\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:12 crc kubenswrapper[4904]: I0214 11:26:12.792637 4904 generic.go:334] "Generic (PLEG): container finished" podID="60a86662-8426-41d7-8eab-bd0ed8ff9fda" containerID="063736057a86ef2533b8d53d185c3959a9c8feed0f08ea11c82cc5ffeeb6c0e0" exitCode=0 Feb 14 11:26:12 crc kubenswrapper[4904]: I0214 11:26:12.792695 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dbdb-account-create-update-6drwj" event={"ID":"60a86662-8426-41d7-8eab-bd0ed8ff9fda","Type":"ContainerDied","Data":"063736057a86ef2533b8d53d185c3959a9c8feed0f08ea11c82cc5ffeeb6c0e0"} Feb 14 11:26:12 crc kubenswrapper[4904]: I0214 11:26:12.793919 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8fad-account-create-update-k8rxj" event={"ID":"a628e9c6-8ff3-41bb-92ae-69d703f224da","Type":"ContainerDied","Data":"0978eb3a719b63da5040d182b949b09be2ebba442474f6fdfddcbb97d695bd81"} Feb 14 11:26:12 crc kubenswrapper[4904]: I0214 11:26:12.793941 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0978eb3a719b63da5040d182b949b09be2ebba442474f6fdfddcbb97d695bd81" Feb 14 11:26:12 crc kubenswrapper[4904]: I0214 11:26:12.793985 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8fad-account-create-update-k8rxj" Feb 14 11:26:12 crc kubenswrapper[4904]: I0214 11:26:12.807608 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0594337d-464f-489f-9761-3d3d2e143af8","Type":"ContainerStarted","Data":"8186f3584b2ea661b3f6be91cea5b426e7c0512df118523648ec064651ec1d8f"} Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.220525 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8v2xr" podUID="a9fd9edc-0058-417c-baa6-9fbff5035af5" containerName="registry-server" probeResult="failure" output=< Feb 14 11:26:13 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 11:26:13 crc kubenswrapper[4904]: > Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.254145 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7m2cz" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.387698 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf6j6\" (UniqueName: \"kubernetes.io/projected/10c49b28-99e8-4e52-9950-bfec689cc77c-kube-api-access-rf6j6\") pod \"10c49b28-99e8-4e52-9950-bfec689cc77c\" (UID: \"10c49b28-99e8-4e52-9950-bfec689cc77c\") " Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.387856 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10c49b28-99e8-4e52-9950-bfec689cc77c-operator-scripts\") pod \"10c49b28-99e8-4e52-9950-bfec689cc77c\" (UID: \"10c49b28-99e8-4e52-9950-bfec689cc77c\") " Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.389512 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10c49b28-99e8-4e52-9950-bfec689cc77c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "10c49b28-99e8-4e52-9950-bfec689cc77c" (UID: "10c49b28-99e8-4e52-9950-bfec689cc77c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.398058 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10c49b28-99e8-4e52-9950-bfec689cc77c-kube-api-access-rf6j6" (OuterVolumeSpecName: "kube-api-access-rf6j6") pod "10c49b28-99e8-4e52-9950-bfec689cc77c" (UID: "10c49b28-99e8-4e52-9950-bfec689cc77c"). InnerVolumeSpecName "kube-api-access-rf6j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.467528 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8b70-account-create-update-j2fhg" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.471150 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vfmnn" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.489867 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf6j6\" (UniqueName: \"kubernetes.io/projected/10c49b28-99e8-4e52-9950-bfec689cc77c-kube-api-access-rf6j6\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.489934 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10c49b28-99e8-4e52-9950-bfec689cc77c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.574495 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9ldqj" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.590670 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4168b6f1-2a3a-410e-a0c6-42783873ab95-operator-scripts\") pod \"4168b6f1-2a3a-410e-a0c6-42783873ab95\" (UID: \"4168b6f1-2a3a-410e-a0c6-42783873ab95\") " Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.590705 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2xb7\" (UniqueName: \"kubernetes.io/projected/4168b6f1-2a3a-410e-a0c6-42783873ab95-kube-api-access-n2xb7\") pod \"4168b6f1-2a3a-410e-a0c6-42783873ab95\" (UID: \"4168b6f1-2a3a-410e-a0c6-42783873ab95\") " Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.590728 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e148cac5-499e-488a-b06b-9aedaa6f2b41-operator-scripts\") pod \"e148cac5-499e-488a-b06b-9aedaa6f2b41\" (UID: \"e148cac5-499e-488a-b06b-9aedaa6f2b41\") " Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.590745 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzrf2\" (UniqueName: \"kubernetes.io/projected/e148cac5-499e-488a-b06b-9aedaa6f2b41-kube-api-access-fzrf2\") pod \"e148cac5-499e-488a-b06b-9aedaa6f2b41\" (UID: \"e148cac5-499e-488a-b06b-9aedaa6f2b41\") " Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.591257 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e148cac5-499e-488a-b06b-9aedaa6f2b41-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e148cac5-499e-488a-b06b-9aedaa6f2b41" (UID: "e148cac5-499e-488a-b06b-9aedaa6f2b41"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.591583 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e148cac5-499e-488a-b06b-9aedaa6f2b41-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.593938 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4168b6f1-2a3a-410e-a0c6-42783873ab95-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4168b6f1-2a3a-410e-a0c6-42783873ab95" (UID: "4168b6f1-2a3a-410e-a0c6-42783873ab95"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.595204 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e148cac5-499e-488a-b06b-9aedaa6f2b41-kube-api-access-fzrf2" (OuterVolumeSpecName: "kube-api-access-fzrf2") pod "e148cac5-499e-488a-b06b-9aedaa6f2b41" (UID: "e148cac5-499e-488a-b06b-9aedaa6f2b41"). InnerVolumeSpecName "kube-api-access-fzrf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.605900 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4168b6f1-2a3a-410e-a0c6-42783873ab95-kube-api-access-n2xb7" (OuterVolumeSpecName: "kube-api-access-n2xb7") pod "4168b6f1-2a3a-410e-a0c6-42783873ab95" (UID: "4168b6f1-2a3a-410e-a0c6-42783873ab95"). InnerVolumeSpecName "kube-api-access-n2xb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.692985 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9215ece4-8699-4e9a-aac3-1c1d2b6822c2-operator-scripts\") pod \"9215ece4-8699-4e9a-aac3-1c1d2b6822c2\" (UID: \"9215ece4-8699-4e9a-aac3-1c1d2b6822c2\") " Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.693177 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgxdd\" (UniqueName: \"kubernetes.io/projected/9215ece4-8699-4e9a-aac3-1c1d2b6822c2-kube-api-access-bgxdd\") pod \"9215ece4-8699-4e9a-aac3-1c1d2b6822c2\" (UID: \"9215ece4-8699-4e9a-aac3-1c1d2b6822c2\") " Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.693501 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4168b6f1-2a3a-410e-a0c6-42783873ab95-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.693520 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2xb7\" (UniqueName: \"kubernetes.io/projected/4168b6f1-2a3a-410e-a0c6-42783873ab95-kube-api-access-n2xb7\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.693531 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzrf2\" (UniqueName: \"kubernetes.io/projected/e148cac5-499e-488a-b06b-9aedaa6f2b41-kube-api-access-fzrf2\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.693543 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9215ece4-8699-4e9a-aac3-1c1d2b6822c2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9215ece4-8699-4e9a-aac3-1c1d2b6822c2" (UID: "9215ece4-8699-4e9a-aac3-1c1d2b6822c2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.696986 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9215ece4-8699-4e9a-aac3-1c1d2b6822c2-kube-api-access-bgxdd" (OuterVolumeSpecName: "kube-api-access-bgxdd") pod "9215ece4-8699-4e9a-aac3-1c1d2b6822c2" (UID: "9215ece4-8699-4e9a-aac3-1c1d2b6822c2"). InnerVolumeSpecName "kube-api-access-bgxdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.795452 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgxdd\" (UniqueName: \"kubernetes.io/projected/9215ece4-8699-4e9a-aac3-1c1d2b6822c2-kube-api-access-bgxdd\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.795484 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9215ece4-8699-4e9a-aac3-1c1d2b6822c2-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.838088 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7m2cz" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.846188 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9ldqj" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.855108 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7m2cz" event={"ID":"10c49b28-99e8-4e52-9950-bfec689cc77c","Type":"ContainerDied","Data":"5d1dd0a60e559a175869ca09c521ad854e2fcf89239960d5db04cd5b9a79a383"} Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.855145 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d1dd0a60e559a175869ca09c521ad854e2fcf89239960d5db04cd5b9a79a383" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.855156 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9ldqj" event={"ID":"9215ece4-8699-4e9a-aac3-1c1d2b6822c2","Type":"ContainerDied","Data":"31b01af65725bed4faed9e4ae06e42f89843d120490d3790d2296c4aae0406d3"} Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.855165 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31b01af65725bed4faed9e4ae06e42f89843d120490d3790d2296c4aae0406d3" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.864647 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vfmnn" event={"ID":"e148cac5-499e-488a-b06b-9aedaa6f2b41","Type":"ContainerDied","Data":"ffd24fa912ddfa84b15920ac561d15d66707308590e726f075a0dd8aca4aa7bb"} Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.864684 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffd24fa912ddfa84b15920ac561d15d66707308590e726f075a0dd8aca4aa7bb" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.864743 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vfmnn" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.872506 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8b70-account-create-update-j2fhg" Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.874230 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8b70-account-create-update-j2fhg" event={"ID":"4168b6f1-2a3a-410e-a0c6-42783873ab95","Type":"ContainerDied","Data":"e467cef0b586f2fa61b2d66de7c2304391effa8f1a3c4b326e129234633c95f8"} Feb 14 11:26:13 crc kubenswrapper[4904]: I0214 11:26:13.880500 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e467cef0b586f2fa61b2d66de7c2304391effa8f1a3c4b326e129234633c95f8" Feb 14 11:26:14 crc kubenswrapper[4904]: I0214 11:26:14.609165 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dbdb-account-create-update-6drwj" Feb 14 11:26:14 crc kubenswrapper[4904]: I0214 11:26:14.720374 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sv8bh\" (UniqueName: \"kubernetes.io/projected/60a86662-8426-41d7-8eab-bd0ed8ff9fda-kube-api-access-sv8bh\") pod \"60a86662-8426-41d7-8eab-bd0ed8ff9fda\" (UID: \"60a86662-8426-41d7-8eab-bd0ed8ff9fda\") " Feb 14 11:26:14 crc kubenswrapper[4904]: I0214 11:26:14.721190 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60a86662-8426-41d7-8eab-bd0ed8ff9fda-operator-scripts\") pod \"60a86662-8426-41d7-8eab-bd0ed8ff9fda\" (UID: \"60a86662-8426-41d7-8eab-bd0ed8ff9fda\") " Feb 14 11:26:14 crc kubenswrapper[4904]: I0214 11:26:14.721775 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60a86662-8426-41d7-8eab-bd0ed8ff9fda-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "60a86662-8426-41d7-8eab-bd0ed8ff9fda" (UID: "60a86662-8426-41d7-8eab-bd0ed8ff9fda"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:14 crc kubenswrapper[4904]: I0214 11:26:14.722148 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60a86662-8426-41d7-8eab-bd0ed8ff9fda-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:14 crc kubenswrapper[4904]: I0214 11:26:14.724924 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60a86662-8426-41d7-8eab-bd0ed8ff9fda-kube-api-access-sv8bh" (OuterVolumeSpecName: "kube-api-access-sv8bh") pod "60a86662-8426-41d7-8eab-bd0ed8ff9fda" (UID: "60a86662-8426-41d7-8eab-bd0ed8ff9fda"). InnerVolumeSpecName "kube-api-access-sv8bh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:14 crc kubenswrapper[4904]: I0214 11:26:14.826049 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sv8bh\" (UniqueName: \"kubernetes.io/projected/60a86662-8426-41d7-8eab-bd0ed8ff9fda-kube-api-access-sv8bh\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:14 crc kubenswrapper[4904]: I0214 11:26:14.900418 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dbdb-account-create-update-6drwj" event={"ID":"60a86662-8426-41d7-8eab-bd0ed8ff9fda","Type":"ContainerDied","Data":"ff27b00163379c4bd4576662061e384aaf59c646dc7b1d4b17d0c032ac02b086"} Feb 14 11:26:14 crc kubenswrapper[4904]: I0214 11:26:14.900458 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff27b00163379c4bd4576662061e384aaf59c646dc7b1d4b17d0c032ac02b086" Feb 14 11:26:14 crc kubenswrapper[4904]: I0214 11:26:14.900515 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dbdb-account-create-update-6drwj" Feb 14 11:26:14 crc kubenswrapper[4904]: I0214 11:26:14.913072 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0594337d-464f-489f-9761-3d3d2e143af8","Type":"ContainerStarted","Data":"1efe6008b9653ad33509124d61175ac8890e4eeaf703d4a3df2b230708d71520"} Feb 14 11:26:16 crc kubenswrapper[4904]: I0214 11:26:16.383116 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:26:16 crc kubenswrapper[4904]: I0214 11:26:16.383469 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:26:16 crc kubenswrapper[4904]: I0214 11:26:16.383514 4904 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:26:16 crc kubenswrapper[4904]: I0214 11:26:16.384322 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fd7271c1e14d135193bc38a4ef3eb1646cd74f3540f5fa874f75aef55d03613d"} pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 14 11:26:16 crc kubenswrapper[4904]: I0214 11:26:16.384384 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" containerID="cri-o://fd7271c1e14d135193bc38a4ef3eb1646cd74f3540f5fa874f75aef55d03613d" gracePeriod=600 Feb 14 11:26:16 crc kubenswrapper[4904]: I0214 11:26:16.932340 4904 generic.go:334] "Generic (PLEG): container finished" podID="7790d276-2506-4641-8c22-67e6441417f0" containerID="fd7271c1e14d135193bc38a4ef3eb1646cd74f3540f5fa874f75aef55d03613d" exitCode=0 Feb 14 11:26:16 crc kubenswrapper[4904]: I0214 11:26:16.932377 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerDied","Data":"fd7271c1e14d135193bc38a4ef3eb1646cd74f3540f5fa874f75aef55d03613d"} Feb 14 11:26:16 crc kubenswrapper[4904]: I0214 11:26:16.932422 4904 scope.go:117] "RemoveContainer" containerID="ed0b33120210f8d03c6d331d3186cbee96523120ba158013d88a03648716b22f" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.005183 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lrm4d"] Feb 14 11:26:18 crc kubenswrapper[4904]: E0214 11:26:18.005702 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e148cac5-499e-488a-b06b-9aedaa6f2b41" containerName="mariadb-database-create" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.005713 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="e148cac5-499e-488a-b06b-9aedaa6f2b41" containerName="mariadb-database-create" Feb 14 11:26:18 crc kubenswrapper[4904]: E0214 11:26:18.005727 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4168b6f1-2a3a-410e-a0c6-42783873ab95" containerName="mariadb-account-create-update" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.005734 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="4168b6f1-2a3a-410e-a0c6-42783873ab95" containerName="mariadb-account-create-update" Feb 14 11:26:18 crc kubenswrapper[4904]: E0214 11:26:18.005747 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60a86662-8426-41d7-8eab-bd0ed8ff9fda" containerName="mariadb-account-create-update" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.005771 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="60a86662-8426-41d7-8eab-bd0ed8ff9fda" containerName="mariadb-account-create-update" Feb 14 11:26:18 crc kubenswrapper[4904]: E0214 11:26:18.005786 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9215ece4-8699-4e9a-aac3-1c1d2b6822c2" containerName="mariadb-database-create" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.005792 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="9215ece4-8699-4e9a-aac3-1c1d2b6822c2" containerName="mariadb-database-create" Feb 14 11:26:18 crc kubenswrapper[4904]: E0214 11:26:18.005805 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73518149-3e47-413e-8bf1-c29279ee89b7" containerName="mariadb-account-create-update" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.005810 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="73518149-3e47-413e-8bf1-c29279ee89b7" containerName="mariadb-account-create-update" Feb 14 11:26:18 crc kubenswrapper[4904]: E0214 11:26:18.005820 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a628e9c6-8ff3-41bb-92ae-69d703f224da" containerName="mariadb-account-create-update" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.005827 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="a628e9c6-8ff3-41bb-92ae-69d703f224da" containerName="mariadb-account-create-update" Feb 14 11:26:18 crc kubenswrapper[4904]: E0214 11:26:18.005851 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10c49b28-99e8-4e52-9950-bfec689cc77c" containerName="mariadb-database-create" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.005857 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="10c49b28-99e8-4e52-9950-bfec689cc77c" containerName="mariadb-database-create" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.006007 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="60a86662-8426-41d7-8eab-bd0ed8ff9fda" containerName="mariadb-account-create-update" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.006017 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="10c49b28-99e8-4e52-9950-bfec689cc77c" containerName="mariadb-database-create" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.006029 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="9215ece4-8699-4e9a-aac3-1c1d2b6822c2" containerName="mariadb-database-create" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.006035 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="73518149-3e47-413e-8bf1-c29279ee89b7" containerName="mariadb-account-create-update" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.006041 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="a628e9c6-8ff3-41bb-92ae-69d703f224da" containerName="mariadb-account-create-update" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.006051 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="4168b6f1-2a3a-410e-a0c6-42783873ab95" containerName="mariadb-account-create-update" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.006065 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="e148cac5-499e-488a-b06b-9aedaa6f2b41" containerName="mariadb-database-create" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.007056 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lrm4d" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.021290 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lrm4d"] Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.185456 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2bcaf28-8b82-460c-9535-b9404a3d19a9-utilities\") pod \"certified-operators-lrm4d\" (UID: \"b2bcaf28-8b82-460c-9535-b9404a3d19a9\") " pod="openshift-marketplace/certified-operators-lrm4d" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.185544 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2bcaf28-8b82-460c-9535-b9404a3d19a9-catalog-content\") pod \"certified-operators-lrm4d\" (UID: \"b2bcaf28-8b82-460c-9535-b9404a3d19a9\") " pod="openshift-marketplace/certified-operators-lrm4d" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.185581 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlkbm\" (UniqueName: \"kubernetes.io/projected/b2bcaf28-8b82-460c-9535-b9404a3d19a9-kube-api-access-jlkbm\") pod \"certified-operators-lrm4d\" (UID: \"b2bcaf28-8b82-460c-9535-b9404a3d19a9\") " pod="openshift-marketplace/certified-operators-lrm4d" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.287118 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2bcaf28-8b82-460c-9535-b9404a3d19a9-utilities\") pod \"certified-operators-lrm4d\" (UID: \"b2bcaf28-8b82-460c-9535-b9404a3d19a9\") " pod="openshift-marketplace/certified-operators-lrm4d" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.287430 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2bcaf28-8b82-460c-9535-b9404a3d19a9-catalog-content\") pod \"certified-operators-lrm4d\" (UID: \"b2bcaf28-8b82-460c-9535-b9404a3d19a9\") " pod="openshift-marketplace/certified-operators-lrm4d" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.287460 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlkbm\" (UniqueName: \"kubernetes.io/projected/b2bcaf28-8b82-460c-9535-b9404a3d19a9-kube-api-access-jlkbm\") pod \"certified-operators-lrm4d\" (UID: \"b2bcaf28-8b82-460c-9535-b9404a3d19a9\") " pod="openshift-marketplace/certified-operators-lrm4d" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.288551 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2bcaf28-8b82-460c-9535-b9404a3d19a9-utilities\") pod \"certified-operators-lrm4d\" (UID: \"b2bcaf28-8b82-460c-9535-b9404a3d19a9\") " pod="openshift-marketplace/certified-operators-lrm4d" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.289025 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2bcaf28-8b82-460c-9535-b9404a3d19a9-catalog-content\") pod \"certified-operators-lrm4d\" (UID: \"b2bcaf28-8b82-460c-9535-b9404a3d19a9\") " pod="openshift-marketplace/certified-operators-lrm4d" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.312166 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlkbm\" (UniqueName: \"kubernetes.io/projected/b2bcaf28-8b82-460c-9535-b9404a3d19a9-kube-api-access-jlkbm\") pod \"certified-operators-lrm4d\" (UID: \"b2bcaf28-8b82-460c-9535-b9404a3d19a9\") " pod="openshift-marketplace/certified-operators-lrm4d" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.328778 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lrm4d" Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.962525 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"d567bf42576edbcc9fce256f34faf5a41303db84f6e9f300d9f0592cbb6a48e2"} Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.967233 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jc69g" event={"ID":"e9a1a642-2b0a-4a5c-825f-883de77215e2","Type":"ContainerStarted","Data":"0e7c0155f80a4151ce218be32053768a5e7f953506fde9f13203f41b15c3c107"} Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.983023 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0594337d-464f-489f-9761-3d3d2e143af8","Type":"ContainerStarted","Data":"b27c7a92fe0b7ed6ea5249cea63ba2f3b54258e8bbe6f7afc3a25fa2ba7bc95f"} Feb 14 11:26:18 crc kubenswrapper[4904]: I0214 11:26:18.983065 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0594337d-464f-489f-9761-3d3d2e143af8","Type":"ContainerStarted","Data":"46e66622fa18ae913d39451a1735d2982c3e9bc717b9e22a49e7974c5926749a"} Feb 14 11:26:19 crc kubenswrapper[4904]: I0214 11:26:19.015480 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-jc69g" podStartSLOduration=3.201819036 podStartE2EDuration="12.01546357s" podCreationTimestamp="2026-02-14 11:26:07 +0000 UTC" firstStartedPulling="2026-02-14 11:26:09.392869486 +0000 UTC m=+960.205634147" lastFinishedPulling="2026-02-14 11:26:18.20651402 +0000 UTC m=+969.019278681" observedRunningTime="2026-02-14 11:26:19.01251579 +0000 UTC m=+969.825280451" watchObservedRunningTime="2026-02-14 11:26:19.01546357 +0000 UTC m=+969.828228231" Feb 14 11:26:19 crc kubenswrapper[4904]: I0214 11:26:19.058935 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lrm4d"] Feb 14 11:26:19 crc kubenswrapper[4904]: W0214 11:26:19.077197 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2bcaf28_8b82_460c_9535_b9404a3d19a9.slice/crio-8f6cf37d10ae646369ca48cc304f7e52812d87fab799e885faa6c4e1a750adfd WatchSource:0}: Error finding container 8f6cf37d10ae646369ca48cc304f7e52812d87fab799e885faa6c4e1a750adfd: Status 404 returned error can't find the container with id 8f6cf37d10ae646369ca48cc304f7e52812d87fab799e885faa6c4e1a750adfd Feb 14 11:26:19 crc kubenswrapper[4904]: I0214 11:26:19.997377 4904 generic.go:334] "Generic (PLEG): container finished" podID="b2bcaf28-8b82-460c-9535-b9404a3d19a9" containerID="f8b1e39b3e8f22c8cbe144a51482aaf198a4fbf2c54ceadf14eaeae193414c3e" exitCode=0 Feb 14 11:26:19 crc kubenswrapper[4904]: I0214 11:26:19.997456 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lrm4d" event={"ID":"b2bcaf28-8b82-460c-9535-b9404a3d19a9","Type":"ContainerDied","Data":"f8b1e39b3e8f22c8cbe144a51482aaf198a4fbf2c54ceadf14eaeae193414c3e"} Feb 14 11:26:19 crc kubenswrapper[4904]: I0214 11:26:19.998033 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lrm4d" event={"ID":"b2bcaf28-8b82-460c-9535-b9404a3d19a9","Type":"ContainerStarted","Data":"8f6cf37d10ae646369ca48cc304f7e52812d87fab799e885faa6c4e1a750adfd"} Feb 14 11:26:20 crc kubenswrapper[4904]: I0214 11:26:20.007319 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0594337d-464f-489f-9761-3d3d2e143af8","Type":"ContainerStarted","Data":"cb65b173541659e56201a5aad064f9021340e8c7caf5fc264d0a42d18e0b6e50"} Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.024492 4904 generic.go:334] "Generic (PLEG): container finished" podID="b2bcaf28-8b82-460c-9535-b9404a3d19a9" containerID="fec6c3df38894bcf3f8752aac31551259a23a5a3fb9b2c2918e0b43dc674e99a" exitCode=0 Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.025383 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lrm4d" event={"ID":"b2bcaf28-8b82-460c-9535-b9404a3d19a9","Type":"ContainerDied","Data":"fec6c3df38894bcf3f8752aac31551259a23a5a3fb9b2c2918e0b43dc674e99a"} Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.030825 4904 generic.go:334] "Generic (PLEG): container finished" podID="042a40bf-ba7e-4f1a-afb9-a4f82a3470e1" containerID="ac18286729f4904e4ad9a3194b4bf19ab0d8831192515607e99661486cdffff4" exitCode=0 Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.030877 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-j4tgk" event={"ID":"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1","Type":"ContainerDied","Data":"ac18286729f4904e4ad9a3194b4bf19ab0d8831192515607e99661486cdffff4"} Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.062389 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0594337d-464f-489f-9761-3d3d2e143af8","Type":"ContainerStarted","Data":"16504cb9203261f2eaf8159ba484f913bb4ba9dbedf74a8d4764c0925c4d4cb2"} Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.062430 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0594337d-464f-489f-9761-3d3d2e143af8","Type":"ContainerStarted","Data":"cfee298220f5053e6d4dd79dfed73fcfa567174f62a0868c46d8393617197f2a"} Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.062438 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0594337d-464f-489f-9761-3d3d2e143af8","Type":"ContainerStarted","Data":"b4b4f3f348649aa0c1d617e126c0e30dd9d7a9ec3b67e2f9a94781b9ac874548"} Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.062446 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0594337d-464f-489f-9761-3d3d2e143af8","Type":"ContainerStarted","Data":"8b41dfa8b8be2ccef130ae19815d4a4d8835f1b44ac25b67662d95192adc505e"} Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.599222 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nbqrf"] Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.601550 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nbqrf" Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.612398 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbqrf"] Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.762219 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6wkq\" (UniqueName: \"kubernetes.io/projected/fd725c16-a4a1-4dca-b7dc-a159d272f4b0-kube-api-access-w6wkq\") pod \"redhat-marketplace-nbqrf\" (UID: \"fd725c16-a4a1-4dca-b7dc-a159d272f4b0\") " pod="openshift-marketplace/redhat-marketplace-nbqrf" Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.762281 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd725c16-a4a1-4dca-b7dc-a159d272f4b0-catalog-content\") pod \"redhat-marketplace-nbqrf\" (UID: \"fd725c16-a4a1-4dca-b7dc-a159d272f4b0\") " pod="openshift-marketplace/redhat-marketplace-nbqrf" Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.762309 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd725c16-a4a1-4dca-b7dc-a159d272f4b0-utilities\") pod \"redhat-marketplace-nbqrf\" (UID: \"fd725c16-a4a1-4dca-b7dc-a159d272f4b0\") " pod="openshift-marketplace/redhat-marketplace-nbqrf" Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.865283 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd725c16-a4a1-4dca-b7dc-a159d272f4b0-utilities\") pod \"redhat-marketplace-nbqrf\" (UID: \"fd725c16-a4a1-4dca-b7dc-a159d272f4b0\") " pod="openshift-marketplace/redhat-marketplace-nbqrf" Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.865694 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6wkq\" (UniqueName: \"kubernetes.io/projected/fd725c16-a4a1-4dca-b7dc-a159d272f4b0-kube-api-access-w6wkq\") pod \"redhat-marketplace-nbqrf\" (UID: \"fd725c16-a4a1-4dca-b7dc-a159d272f4b0\") " pod="openshift-marketplace/redhat-marketplace-nbqrf" Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.865733 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd725c16-a4a1-4dca-b7dc-a159d272f4b0-catalog-content\") pod \"redhat-marketplace-nbqrf\" (UID: \"fd725c16-a4a1-4dca-b7dc-a159d272f4b0\") " pod="openshift-marketplace/redhat-marketplace-nbqrf" Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.866788 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd725c16-a4a1-4dca-b7dc-a159d272f4b0-utilities\") pod \"redhat-marketplace-nbqrf\" (UID: \"fd725c16-a4a1-4dca-b7dc-a159d272f4b0\") " pod="openshift-marketplace/redhat-marketplace-nbqrf" Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.866919 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd725c16-a4a1-4dca-b7dc-a159d272f4b0-catalog-content\") pod \"redhat-marketplace-nbqrf\" (UID: \"fd725c16-a4a1-4dca-b7dc-a159d272f4b0\") " pod="openshift-marketplace/redhat-marketplace-nbqrf" Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.886096 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6wkq\" (UniqueName: \"kubernetes.io/projected/fd725c16-a4a1-4dca-b7dc-a159d272f4b0-kube-api-access-w6wkq\") pod \"redhat-marketplace-nbqrf\" (UID: \"fd725c16-a4a1-4dca-b7dc-a159d272f4b0\") " pod="openshift-marketplace/redhat-marketplace-nbqrf" Feb 14 11:26:22 crc kubenswrapper[4904]: I0214 11:26:22.916891 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nbqrf" Feb 14 11:26:23 crc kubenswrapper[4904]: I0214 11:26:23.114516 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0594337d-464f-489f-9761-3d3d2e143af8","Type":"ContainerStarted","Data":"631cdd3aef938645ca6f04dd74e27b31d8cdc91c64f581bced9026aa90a33543"} Feb 14 11:26:23 crc kubenswrapper[4904]: I0214 11:26:23.115344 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0594337d-464f-489f-9761-3d3d2e143af8","Type":"ContainerStarted","Data":"16b59a892da14c9a1159284310d91bfa18f64636eb74ac9f2dd51573caba8cb9"} Feb 14 11:26:23 crc kubenswrapper[4904]: I0214 11:26:23.213429 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8v2xr" podUID="a9fd9edc-0058-417c-baa6-9fbff5035af5" containerName="registry-server" probeResult="failure" output=< Feb 14 11:26:23 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 11:26:23 crc kubenswrapper[4904]: > Feb 14 11:26:23 crc kubenswrapper[4904]: I0214 11:26:23.425574 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbqrf"] Feb 14 11:26:23 crc kubenswrapper[4904]: I0214 11:26:23.630420 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-j4tgk" Feb 14 11:26:23 crc kubenswrapper[4904]: I0214 11:26:23.683533 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-db-sync-config-data\") pod \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\" (UID: \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\") " Feb 14 11:26:23 crc kubenswrapper[4904]: I0214 11:26:23.683654 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-combined-ca-bundle\") pod \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\" (UID: \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\") " Feb 14 11:26:23 crc kubenswrapper[4904]: I0214 11:26:23.683677 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-config-data\") pod \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\" (UID: \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\") " Feb 14 11:26:23 crc kubenswrapper[4904]: I0214 11:26:23.683708 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdgrf\" (UniqueName: \"kubernetes.io/projected/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-kube-api-access-tdgrf\") pod \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\" (UID: \"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1\") " Feb 14 11:26:23 crc kubenswrapper[4904]: I0214 11:26:23.693414 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "042a40bf-ba7e-4f1a-afb9-a4f82a3470e1" (UID: "042a40bf-ba7e-4f1a-afb9-a4f82a3470e1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:23 crc kubenswrapper[4904]: I0214 11:26:23.693826 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-kube-api-access-tdgrf" (OuterVolumeSpecName: "kube-api-access-tdgrf") pod "042a40bf-ba7e-4f1a-afb9-a4f82a3470e1" (UID: "042a40bf-ba7e-4f1a-afb9-a4f82a3470e1"). InnerVolumeSpecName "kube-api-access-tdgrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:23 crc kubenswrapper[4904]: I0214 11:26:23.733669 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "042a40bf-ba7e-4f1a-afb9-a4f82a3470e1" (UID: "042a40bf-ba7e-4f1a-afb9-a4f82a3470e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:23 crc kubenswrapper[4904]: I0214 11:26:23.761093 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-config-data" (OuterVolumeSpecName: "config-data") pod "042a40bf-ba7e-4f1a-afb9-a4f82a3470e1" (UID: "042a40bf-ba7e-4f1a-afb9-a4f82a3470e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:23 crc kubenswrapper[4904]: I0214 11:26:23.785321 4904 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:23 crc kubenswrapper[4904]: I0214 11:26:23.785361 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:23 crc kubenswrapper[4904]: I0214 11:26:23.785371 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:23 crc kubenswrapper[4904]: I0214 11:26:23.785381 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdgrf\" (UniqueName: \"kubernetes.io/projected/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1-kube-api-access-tdgrf\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.130689 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0594337d-464f-489f-9761-3d3d2e143af8","Type":"ContainerStarted","Data":"da18ad6fa7c03ebf8fdee9f02d069d4bf60960466707a118bc2c76d9f5ea5010"} Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.132344 4904 generic.go:334] "Generic (PLEG): container finished" podID="fd725c16-a4a1-4dca-b7dc-a159d272f4b0" containerID="200a3f4b00beaba2c6572617b88e7d9657462037d59c847932ce70f8e1b693c3" exitCode=0 Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.132396 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbqrf" event={"ID":"fd725c16-a4a1-4dca-b7dc-a159d272f4b0","Type":"ContainerDied","Data":"200a3f4b00beaba2c6572617b88e7d9657462037d59c847932ce70f8e1b693c3"} Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.132412 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbqrf" event={"ID":"fd725c16-a4a1-4dca-b7dc-a159d272f4b0","Type":"ContainerStarted","Data":"9c03e5b377a38f43a5317a99e4ddb964e518e6f2e15dbf8f7c0e1e1d37fbd5ae"} Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.134619 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lrm4d" event={"ID":"b2bcaf28-8b82-460c-9535-b9404a3d19a9","Type":"ContainerStarted","Data":"db123fad5e9b0e82052c07b64897a7abfc8be7140e0070fcf018b76330f9cb21"} Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.137223 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-j4tgk" event={"ID":"042a40bf-ba7e-4f1a-afb9-a4f82a3470e1","Type":"ContainerDied","Data":"17fc667c28b358c070356f3e1967adedff6b9e1d79be4711e87f5ac328d099b4"} Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.137253 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17fc667c28b358c070356f3e1967adedff6b9e1d79be4711e87f5ac328d099b4" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.137367 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-j4tgk" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.174554 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=37.60375303 podStartE2EDuration="52.174536513s" podCreationTimestamp="2026-02-14 11:25:32 +0000 UTC" firstStartedPulling="2026-02-14 11:26:06.404059059 +0000 UTC m=+957.216823720" lastFinishedPulling="2026-02-14 11:26:20.974842552 +0000 UTC m=+971.787607203" observedRunningTime="2026-02-14 11:26:24.169265518 +0000 UTC m=+974.982030169" watchObservedRunningTime="2026-02-14 11:26:24.174536513 +0000 UTC m=+974.987301174" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.230916 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lrm4d" podStartSLOduration=4.443169573 podStartE2EDuration="7.230895488s" podCreationTimestamp="2026-02-14 11:26:17 +0000 UTC" firstStartedPulling="2026-02-14 11:26:19.999253684 +0000 UTC m=+970.812018355" lastFinishedPulling="2026-02-14 11:26:22.786979619 +0000 UTC m=+973.599744270" observedRunningTime="2026-02-14 11:26:24.224930345 +0000 UTC m=+975.037695006" watchObservedRunningTime="2026-02-14 11:26:24.230895488 +0000 UTC m=+975.043660149" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.648594 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-zhfbz"] Feb 14 11:26:24 crc kubenswrapper[4904]: E0214 11:26:24.649253 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="042a40bf-ba7e-4f1a-afb9-a4f82a3470e1" containerName="glance-db-sync" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.649268 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="042a40bf-ba7e-4f1a-afb9-a4f82a3470e1" containerName="glance-db-sync" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.649439 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="042a40bf-ba7e-4f1a-afb9-a4f82a3470e1" containerName="glance-db-sync" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.650236 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.662522 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-zhfbz"] Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.701203 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgccj\" (UniqueName: \"kubernetes.io/projected/626e3a44-2cab-4e42-8693-3ecc3bd34652-kube-api-access-hgccj\") pod \"dnsmasq-dns-5b946c75cc-zhfbz\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.701352 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-zhfbz\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.701394 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-zhfbz\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.701414 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-zhfbz\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.701459 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-config\") pod \"dnsmasq-dns-5b946c75cc-zhfbz\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.785710 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-zhfbz"] Feb 14 11:26:24 crc kubenswrapper[4904]: E0214 11:26:24.786282 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-hgccj ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" podUID="626e3a44-2cab-4e42-8693-3ecc3bd34652" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.802878 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-zhfbz\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.802977 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-zhfbz\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.803035 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-config\") pod \"dnsmasq-dns-5b946c75cc-zhfbz\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.803072 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgccj\" (UniqueName: \"kubernetes.io/projected/626e3a44-2cab-4e42-8693-3ecc3bd34652-kube-api-access-hgccj\") pod \"dnsmasq-dns-5b946c75cc-zhfbz\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.803160 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-zhfbz\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.803693 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-zhfbz\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.804882 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-zhfbz\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.805430 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-config\") pod \"dnsmasq-dns-5b946c75cc-zhfbz\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.806267 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-zhfbz\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.829627 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-lrs72"] Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.835626 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.846103 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.859225 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgccj\" (UniqueName: \"kubernetes.io/projected/626e3a44-2cab-4e42-8693-3ecc3bd34652-kube-api-access-hgccj\") pod \"dnsmasq-dns-5b946c75cc-zhfbz\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" Feb 14 11:26:24 crc kubenswrapper[4904]: I0214 11:26:24.879129 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-lrs72"] Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.005652 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-lrs72\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.005991 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-lrs72\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.006038 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-lrs72\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.006075 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hntdn\" (UniqueName: \"kubernetes.io/projected/916f82bd-6f92-4e00-be47-d3343219d381-kube-api-access-hntdn\") pod \"dnsmasq-dns-74f6bcbc87-lrs72\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.006105 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-lrs72\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.006136 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-config\") pod \"dnsmasq-dns-74f6bcbc87-lrs72\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.107766 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-lrs72\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.107825 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-lrs72\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.107875 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-lrs72\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.107913 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hntdn\" (UniqueName: \"kubernetes.io/projected/916f82bd-6f92-4e00-be47-d3343219d381-kube-api-access-hntdn\") pod \"dnsmasq-dns-74f6bcbc87-lrs72\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.107944 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-lrs72\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.107974 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-config\") pod \"dnsmasq-dns-74f6bcbc87-lrs72\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.108737 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-lrs72\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.108754 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-lrs72\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.108754 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-lrs72\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.108865 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-config\") pod \"dnsmasq-dns-74f6bcbc87-lrs72\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.109310 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-lrs72\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.134911 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hntdn\" (UniqueName: \"kubernetes.io/projected/916f82bd-6f92-4e00-be47-d3343219d381-kube-api-access-hntdn\") pod \"dnsmasq-dns-74f6bcbc87-lrs72\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.149570 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.180308 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.247054 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.311103 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-ovsdbserver-nb\") pod \"626e3a44-2cab-4e42-8693-3ecc3bd34652\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.311228 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgccj\" (UniqueName: \"kubernetes.io/projected/626e3a44-2cab-4e42-8693-3ecc3bd34652-kube-api-access-hgccj\") pod \"626e3a44-2cab-4e42-8693-3ecc3bd34652\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.311278 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-config\") pod \"626e3a44-2cab-4e42-8693-3ecc3bd34652\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.311362 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-dns-svc\") pod \"626e3a44-2cab-4e42-8693-3ecc3bd34652\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.311430 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-ovsdbserver-sb\") pod \"626e3a44-2cab-4e42-8693-3ecc3bd34652\" (UID: \"626e3a44-2cab-4e42-8693-3ecc3bd34652\") " Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.311481 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "626e3a44-2cab-4e42-8693-3ecc3bd34652" (UID: "626e3a44-2cab-4e42-8693-3ecc3bd34652"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.311920 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "626e3a44-2cab-4e42-8693-3ecc3bd34652" (UID: "626e3a44-2cab-4e42-8693-3ecc3bd34652"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.311935 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-config" (OuterVolumeSpecName: "config") pod "626e3a44-2cab-4e42-8693-3ecc3bd34652" (UID: "626e3a44-2cab-4e42-8693-3ecc3bd34652"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.312093 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "626e3a44-2cab-4e42-8693-3ecc3bd34652" (UID: "626e3a44-2cab-4e42-8693-3ecc3bd34652"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.312204 4904 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.312221 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.312232 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.312241 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/626e3a44-2cab-4e42-8693-3ecc3bd34652-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.319817 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/626e3a44-2cab-4e42-8693-3ecc3bd34652-kube-api-access-hgccj" (OuterVolumeSpecName: "kube-api-access-hgccj") pod "626e3a44-2cab-4e42-8693-3ecc3bd34652" (UID: "626e3a44-2cab-4e42-8693-3ecc3bd34652"). InnerVolumeSpecName "kube-api-access-hgccj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.431242 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgccj\" (UniqueName: \"kubernetes.io/projected/626e3a44-2cab-4e42-8693-3ecc3bd34652-kube-api-access-hgccj\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:25 crc kubenswrapper[4904]: I0214 11:26:25.739908 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-lrs72"] Feb 14 11:26:25 crc kubenswrapper[4904]: W0214 11:26:25.749125 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod916f82bd_6f92_4e00_be47_d3343219d381.slice/crio-3cbc0cee9bb5d92e61e8983e7bcb2a6485fc88156a858f8bc209e98b9d39d58a WatchSource:0}: Error finding container 3cbc0cee9bb5d92e61e8983e7bcb2a6485fc88156a858f8bc209e98b9d39d58a: Status 404 returned error can't find the container with id 3cbc0cee9bb5d92e61e8983e7bcb2a6485fc88156a858f8bc209e98b9d39d58a Feb 14 11:26:26 crc kubenswrapper[4904]: I0214 11:26:26.163393 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" event={"ID":"916f82bd-6f92-4e00-be47-d3343219d381","Type":"ContainerStarted","Data":"ef1ad53d02639032e085934ce28fe8a2f0f54d525265966d6d22e0c72e12a9c7"} Feb 14 11:26:26 crc kubenswrapper[4904]: I0214 11:26:26.163691 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" event={"ID":"916f82bd-6f92-4e00-be47-d3343219d381","Type":"ContainerStarted","Data":"3cbc0cee9bb5d92e61e8983e7bcb2a6485fc88156a858f8bc209e98b9d39d58a"} Feb 14 11:26:26 crc kubenswrapper[4904]: I0214 11:26:26.166887 4904 generic.go:334] "Generic (PLEG): container finished" podID="e9a1a642-2b0a-4a5c-825f-883de77215e2" containerID="0e7c0155f80a4151ce218be32053768a5e7f953506fde9f13203f41b15c3c107" exitCode=0 Feb 14 11:26:26 crc kubenswrapper[4904]: I0214 11:26:26.166981 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jc69g" event={"ID":"e9a1a642-2b0a-4a5c-825f-883de77215e2","Type":"ContainerDied","Data":"0e7c0155f80a4151ce218be32053768a5e7f953506fde9f13203f41b15c3c107"} Feb 14 11:26:26 crc kubenswrapper[4904]: I0214 11:26:26.169768 4904 generic.go:334] "Generic (PLEG): container finished" podID="fd725c16-a4a1-4dca-b7dc-a159d272f4b0" containerID="0d444ca3e2b3c8d4fd3da4f79b5df977d653d191a3fff863c754d792a27f0d5e" exitCode=0 Feb 14 11:26:26 crc kubenswrapper[4904]: I0214 11:26:26.169861 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-zhfbz" Feb 14 11:26:26 crc kubenswrapper[4904]: I0214 11:26:26.169972 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbqrf" event={"ID":"fd725c16-a4a1-4dca-b7dc-a159d272f4b0","Type":"ContainerDied","Data":"0d444ca3e2b3c8d4fd3da4f79b5df977d653d191a3fff863c754d792a27f0d5e"} Feb 14 11:26:26 crc kubenswrapper[4904]: I0214 11:26:26.263053 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-zhfbz"] Feb 14 11:26:26 crc kubenswrapper[4904]: I0214 11:26:26.265477 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-zhfbz"] Feb 14 11:26:27 crc kubenswrapper[4904]: I0214 11:26:27.178884 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbqrf" event={"ID":"fd725c16-a4a1-4dca-b7dc-a159d272f4b0","Type":"ContainerStarted","Data":"58f134133db6d54efd6509fd24fa849311a43dc0c4a199b39e5681da001f7985"} Feb 14 11:26:27 crc kubenswrapper[4904]: I0214 11:26:27.180238 4904 generic.go:334] "Generic (PLEG): container finished" podID="916f82bd-6f92-4e00-be47-d3343219d381" containerID="ef1ad53d02639032e085934ce28fe8a2f0f54d525265966d6d22e0c72e12a9c7" exitCode=0 Feb 14 11:26:27 crc kubenswrapper[4904]: I0214 11:26:27.180446 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" event={"ID":"916f82bd-6f92-4e00-be47-d3343219d381","Type":"ContainerDied","Data":"ef1ad53d02639032e085934ce28fe8a2f0f54d525265966d6d22e0c72e12a9c7"} Feb 14 11:26:27 crc kubenswrapper[4904]: I0214 11:26:27.235297 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nbqrf" podStartSLOduration=2.815803006 podStartE2EDuration="5.235281793s" podCreationTimestamp="2026-02-14 11:26:22 +0000 UTC" firstStartedPulling="2026-02-14 11:26:24.134279289 +0000 UTC m=+974.947043960" lastFinishedPulling="2026-02-14 11:26:26.553758086 +0000 UTC m=+977.366522747" observedRunningTime="2026-02-14 11:26:27.22645687 +0000 UTC m=+978.039221531" watchObservedRunningTime="2026-02-14 11:26:27.235281793 +0000 UTC m=+978.048046454" Feb 14 11:26:27 crc kubenswrapper[4904]: I0214 11:26:27.597678 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jc69g" Feb 14 11:26:27 crc kubenswrapper[4904]: I0214 11:26:27.773689 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a1a642-2b0a-4a5c-825f-883de77215e2-combined-ca-bundle\") pod \"e9a1a642-2b0a-4a5c-825f-883de77215e2\" (UID: \"e9a1a642-2b0a-4a5c-825f-883de77215e2\") " Feb 14 11:26:27 crc kubenswrapper[4904]: I0214 11:26:27.774058 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6fzz\" (UniqueName: \"kubernetes.io/projected/e9a1a642-2b0a-4a5c-825f-883de77215e2-kube-api-access-v6fzz\") pod \"e9a1a642-2b0a-4a5c-825f-883de77215e2\" (UID: \"e9a1a642-2b0a-4a5c-825f-883de77215e2\") " Feb 14 11:26:27 crc kubenswrapper[4904]: I0214 11:26:27.774106 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9a1a642-2b0a-4a5c-825f-883de77215e2-config-data\") pod \"e9a1a642-2b0a-4a5c-825f-883de77215e2\" (UID: \"e9a1a642-2b0a-4a5c-825f-883de77215e2\") " Feb 14 11:26:27 crc kubenswrapper[4904]: I0214 11:26:27.779653 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9a1a642-2b0a-4a5c-825f-883de77215e2-kube-api-access-v6fzz" (OuterVolumeSpecName: "kube-api-access-v6fzz") pod "e9a1a642-2b0a-4a5c-825f-883de77215e2" (UID: "e9a1a642-2b0a-4a5c-825f-883de77215e2"). InnerVolumeSpecName "kube-api-access-v6fzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:27 crc kubenswrapper[4904]: I0214 11:26:27.805978 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9a1a642-2b0a-4a5c-825f-883de77215e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e9a1a642-2b0a-4a5c-825f-883de77215e2" (UID: "e9a1a642-2b0a-4a5c-825f-883de77215e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:27 crc kubenswrapper[4904]: I0214 11:26:27.824934 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9a1a642-2b0a-4a5c-825f-883de77215e2-config-data" (OuterVolumeSpecName: "config-data") pod "e9a1a642-2b0a-4a5c-825f-883de77215e2" (UID: "e9a1a642-2b0a-4a5c-825f-883de77215e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:27 crc kubenswrapper[4904]: I0214 11:26:27.846597 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="626e3a44-2cab-4e42-8693-3ecc3bd34652" path="/var/lib/kubelet/pods/626e3a44-2cab-4e42-8693-3ecc3bd34652/volumes" Feb 14 11:26:27 crc kubenswrapper[4904]: I0214 11:26:27.875388 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6fzz\" (UniqueName: \"kubernetes.io/projected/e9a1a642-2b0a-4a5c-825f-883de77215e2-kube-api-access-v6fzz\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:27 crc kubenswrapper[4904]: I0214 11:26:27.875432 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9a1a642-2b0a-4a5c-825f-883de77215e2-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:27 crc kubenswrapper[4904]: I0214 11:26:27.875444 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a1a642-2b0a-4a5c-825f-883de77215e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.188384 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jc69g" event={"ID":"e9a1a642-2b0a-4a5c-825f-883de77215e2","Type":"ContainerDied","Data":"58d8d4fd4d8d1c3014dda653bd847d5e9088534b77ab149a1dec761647276cf2"} Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.188418 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58d8d4fd4d8d1c3014dda653bd847d5e9088534b77ab149a1dec761647276cf2" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.188472 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jc69g" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.191698 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" event={"ID":"916f82bd-6f92-4e00-be47-d3343219d381","Type":"ContainerStarted","Data":"ba6f63f0f7658c40d32342c79fd211328b0e6234523c5ad93b1122e4bf50d469"} Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.219696 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" podStartSLOduration=4.2196759329999995 podStartE2EDuration="4.219675933s" podCreationTimestamp="2026-02-14 11:26:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:26:28.218797289 +0000 UTC m=+979.031561950" watchObservedRunningTime="2026-02-14 11:26:28.219675933 +0000 UTC m=+979.032440604" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.328972 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lrm4d" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.329337 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lrm4d" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.505528 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-lrs72"] Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.564730 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-7bjcl"] Feb 14 11:26:28 crc kubenswrapper[4904]: E0214 11:26:28.565344 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9a1a642-2b0a-4a5c-825f-883de77215e2" containerName="keystone-db-sync" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.565414 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9a1a642-2b0a-4a5c-825f-883de77215e2" containerName="keystone-db-sync" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.565617 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9a1a642-2b0a-4a5c-825f-883de77215e2" containerName="keystone-db-sync" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.568625 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.576799 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-dxj9p"] Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.576999 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.577164 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 14 11:26:28 crc kubenswrapper[4904]: W0214 11:26:28.577331 4904 reflector.go:561] object-"openstack"/"osp-secret": failed to list *v1.Secret: secrets "osp-secret" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Feb 14 11:26:28 crc kubenswrapper[4904]: E0214 11:26:28.577366 4904 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"osp-secret\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"osp-secret\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.577426 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.577685 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-57x7t" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.591060 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.596319 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-scripts\") pod \"keystone-bootstrap-7bjcl\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.596365 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpmc2\" (UniqueName: \"kubernetes.io/projected/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-kube-api-access-tpmc2\") pod \"dnsmasq-dns-847c4cc679-dxj9p\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.596387 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-credential-keys\") pod \"keystone-bootstrap-7bjcl\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.596424 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-dxj9p\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.596445 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-config\") pod \"dnsmasq-dns-847c4cc679-dxj9p\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.596472 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-combined-ca-bundle\") pod \"keystone-bootstrap-7bjcl\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.596487 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-dxj9p\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.596516 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-dxj9p\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.596542 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkvld\" (UniqueName: \"kubernetes.io/projected/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-kube-api-access-qkvld\") pod \"keystone-bootstrap-7bjcl\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.596576 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-config-data\") pod \"keystone-bootstrap-7bjcl\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.596590 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-dns-svc\") pod \"dnsmasq-dns-847c4cc679-dxj9p\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.596621 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-fernet-keys\") pod \"keystone-bootstrap-7bjcl\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.598862 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-7bjcl"] Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.643400 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-dxj9p"] Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.697681 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-scripts\") pod \"keystone-bootstrap-7bjcl\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.697728 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpmc2\" (UniqueName: \"kubernetes.io/projected/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-kube-api-access-tpmc2\") pod \"dnsmasq-dns-847c4cc679-dxj9p\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.697749 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-credential-keys\") pod \"keystone-bootstrap-7bjcl\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.697765 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-dxj9p\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.697783 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-config\") pod \"dnsmasq-dns-847c4cc679-dxj9p\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.697814 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-combined-ca-bundle\") pod \"keystone-bootstrap-7bjcl\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.697847 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-dxj9p\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.697869 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-dxj9p\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.697896 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkvld\" (UniqueName: \"kubernetes.io/projected/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-kube-api-access-qkvld\") pod \"keystone-bootstrap-7bjcl\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.697929 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-config-data\") pod \"keystone-bootstrap-7bjcl\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.697942 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-dns-svc\") pod \"dnsmasq-dns-847c4cc679-dxj9p\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.697972 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-fernet-keys\") pod \"keystone-bootstrap-7bjcl\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.703415 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-dxj9p\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.704573 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-dxj9p\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.707855 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-fernet-keys\") pod \"keystone-bootstrap-7bjcl\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.708660 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-dns-svc\") pod \"dnsmasq-dns-847c4cc679-dxj9p\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.710638 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-dxj9p\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.711607 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-config\") pod \"dnsmasq-dns-847c4cc679-dxj9p\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.712659 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-combined-ca-bundle\") pod \"keystone-bootstrap-7bjcl\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.713139 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-credential-keys\") pod \"keystone-bootstrap-7bjcl\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.724666 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-config-data\") pod \"keystone-bootstrap-7bjcl\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.730229 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-scripts\") pod \"keystone-bootstrap-7bjcl\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.763279 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpmc2\" (UniqueName: \"kubernetes.io/projected/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-kube-api-access-tpmc2\") pod \"dnsmasq-dns-847c4cc679-dxj9p\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.782445 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkvld\" (UniqueName: \"kubernetes.io/projected/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-kube-api-access-qkvld\") pod \"keystone-bootstrap-7bjcl\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.892235 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.912215 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.912667 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6b79d987b9-h8nmk"] Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.913944 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b79d987b9-h8nmk" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.924852 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbx78\" (UniqueName: \"kubernetes.io/projected/97bc1e53-5edf-43fc-8aca-0dee904c86eb-kube-api-access-tbx78\") pod \"horizon-6b79d987b9-h8nmk\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " pod="openstack/horizon-6b79d987b9-h8nmk" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.924907 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/97bc1e53-5edf-43fc-8aca-0dee904c86eb-horizon-secret-key\") pod \"horizon-6b79d987b9-h8nmk\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " pod="openstack/horizon-6b79d987b9-h8nmk" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.924947 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97bc1e53-5edf-43fc-8aca-0dee904c86eb-scripts\") pod \"horizon-6b79d987b9-h8nmk\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " pod="openstack/horizon-6b79d987b9-h8nmk" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.924999 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97bc1e53-5edf-43fc-8aca-0dee904c86eb-logs\") pod \"horizon-6b79d987b9-h8nmk\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " pod="openstack/horizon-6b79d987b9-h8nmk" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.925018 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97bc1e53-5edf-43fc-8aca-0dee904c86eb-config-data\") pod \"horizon-6b79d987b9-h8nmk\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " pod="openstack/horizon-6b79d987b9-h8nmk" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.930590 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.930953 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.931161 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-gnskb" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.931309 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Feb 14 11:26:28 crc kubenswrapper[4904]: I0214 11:26:28.972442 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6b79d987b9-h8nmk"] Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.012781 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-t6bn6"] Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.013902 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-t6bn6" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.026430 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbx78\" (UniqueName: \"kubernetes.io/projected/97bc1e53-5edf-43fc-8aca-0dee904c86eb-kube-api-access-tbx78\") pod \"horizon-6b79d987b9-h8nmk\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " pod="openstack/horizon-6b79d987b9-h8nmk" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.026533 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/97bc1e53-5edf-43fc-8aca-0dee904c86eb-horizon-secret-key\") pod \"horizon-6b79d987b9-h8nmk\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " pod="openstack/horizon-6b79d987b9-h8nmk" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.026629 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97bc1e53-5edf-43fc-8aca-0dee904c86eb-scripts\") pod \"horizon-6b79d987b9-h8nmk\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " pod="openstack/horizon-6b79d987b9-h8nmk" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.026733 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97bc1e53-5edf-43fc-8aca-0dee904c86eb-logs\") pod \"horizon-6b79d987b9-h8nmk\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " pod="openstack/horizon-6b79d987b9-h8nmk" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.026805 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97bc1e53-5edf-43fc-8aca-0dee904c86eb-config-data\") pod \"horizon-6b79d987b9-h8nmk\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " pod="openstack/horizon-6b79d987b9-h8nmk" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.028520 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97bc1e53-5edf-43fc-8aca-0dee904c86eb-config-data\") pod \"horizon-6b79d987b9-h8nmk\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " pod="openstack/horizon-6b79d987b9-h8nmk" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.029107 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97bc1e53-5edf-43fc-8aca-0dee904c86eb-scripts\") pod \"horizon-6b79d987b9-h8nmk\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " pod="openstack/horizon-6b79d987b9-h8nmk" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.029402 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97bc1e53-5edf-43fc-8aca-0dee904c86eb-logs\") pod \"horizon-6b79d987b9-h8nmk\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " pod="openstack/horizon-6b79d987b9-h8nmk" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.037903 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.038360 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.038552 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-rjj2h" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.040527 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/97bc1e53-5edf-43fc-8aca-0dee904c86eb-horizon-secret-key\") pod \"horizon-6b79d987b9-h8nmk\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " pod="openstack/horizon-6b79d987b9-h8nmk" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.087065 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbx78\" (UniqueName: \"kubernetes.io/projected/97bc1e53-5edf-43fc-8aca-0dee904c86eb-kube-api-access-tbx78\") pod \"horizon-6b79d987b9-h8nmk\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " pod="openstack/horizon-6b79d987b9-h8nmk" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.094899 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-t6bn6"] Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.130896 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2abcf43f-6f3b-4804-83a5-ba42fc66f2ec-combined-ca-bundle\") pod \"neutron-db-sync-t6bn6\" (UID: \"2abcf43f-6f3b-4804-83a5-ba42fc66f2ec\") " pod="openstack/neutron-db-sync-t6bn6" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.130975 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2abcf43f-6f3b-4804-83a5-ba42fc66f2ec-config\") pod \"neutron-db-sync-t6bn6\" (UID: \"2abcf43f-6f3b-4804-83a5-ba42fc66f2ec\") " pod="openstack/neutron-db-sync-t6bn6" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.131037 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzt7z\" (UniqueName: \"kubernetes.io/projected/2abcf43f-6f3b-4804-83a5-ba42fc66f2ec-kube-api-access-wzt7z\") pod \"neutron-db-sync-t6bn6\" (UID: \"2abcf43f-6f3b-4804-83a5-ba42fc66f2ec\") " pod="openstack/neutron-db-sync-t6bn6" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.134046 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-dxj9p"] Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.170310 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.172186 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.177297 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.177524 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.202073 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.222598 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.233158 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzt7z\" (UniqueName: \"kubernetes.io/projected/2abcf43f-6f3b-4804-83a5-ba42fc66f2ec-kube-api-access-wzt7z\") pod \"neutron-db-sync-t6bn6\" (UID: \"2abcf43f-6f3b-4804-83a5-ba42fc66f2ec\") " pod="openstack/neutron-db-sync-t6bn6" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.233282 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2abcf43f-6f3b-4804-83a5-ba42fc66f2ec-combined-ca-bundle\") pod \"neutron-db-sync-t6bn6\" (UID: \"2abcf43f-6f3b-4804-83a5-ba42fc66f2ec\") " pod="openstack/neutron-db-sync-t6bn6" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.233353 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2abcf43f-6f3b-4804-83a5-ba42fc66f2ec-config\") pod \"neutron-db-sync-t6bn6\" (UID: \"2abcf43f-6f3b-4804-83a5-ba42fc66f2ec\") " pod="openstack/neutron-db-sync-t6bn6" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.242240 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2abcf43f-6f3b-4804-83a5-ba42fc66f2ec-combined-ca-bundle\") pod \"neutron-db-sync-t6bn6\" (UID: \"2abcf43f-6f3b-4804-83a5-ba42fc66f2ec\") " pod="openstack/neutron-db-sync-t6bn6" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.246628 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2abcf43f-6f3b-4804-83a5-ba42fc66f2ec-config\") pod \"neutron-db-sync-t6bn6\" (UID: \"2abcf43f-6f3b-4804-83a5-ba42fc66f2ec\") " pod="openstack/neutron-db-sync-t6bn6" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.253081 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-nnlq9"] Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.254124 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.254293 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-5wt4c"] Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.280578 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b79d987b9-h8nmk" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.282162 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-rvw8l" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.282425 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.282482 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.282785 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.329941 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-874bc5bd9-jqhms"] Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.331350 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-874bc5bd9-jqhms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.337572 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-5wt4c"] Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.338418 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16339330-a7b5-471a-ad7e-1a8d4627c3e2-log-httpd\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.338460 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16339330-a7b5-471a-ad7e-1a8d4627c3e2-run-httpd\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.338517 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.338584 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-config-data\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.338621 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-scripts\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.338639 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgmqf\" (UniqueName: \"kubernetes.io/projected/16339330-a7b5-471a-ad7e-1a8d4627c3e2-kube-api-access-xgmqf\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.338712 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.361673 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-nnlq9"] Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.387941 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzt7z\" (UniqueName: \"kubernetes.io/projected/2abcf43f-6f3b-4804-83a5-ba42fc66f2ec-kube-api-access-wzt7z\") pod \"neutron-db-sync-t6bn6\" (UID: \"2abcf43f-6f3b-4804-83a5-ba42fc66f2ec\") " pod="openstack/neutron-db-sync-t6bn6" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.388012 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-ft76h"] Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.389331 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ft76h" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.402178 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-ndvl8" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.402302 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-lrm4d" podUID="b2bcaf28-8b82-460c-9535-b9404a3d19a9" containerName="registry-server" probeResult="failure" output=< Feb 14 11:26:29 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 11:26:29 crc kubenswrapper[4904]: > Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.402366 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.402477 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.402620 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-sflms"] Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.403654 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sflms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.409098 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.409259 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-glj86" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.436969 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-sflms"] Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.488688 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c698a36a-d049-4fdb-b279-1b62d1f807d3-etc-machine-id\") pod \"cinder-db-sync-nnlq9\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.490468 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-config-data\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.490567 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-scripts\") pod \"cinder-db-sync-nnlq9\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.490675 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-scripts\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.490778 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-5wt4c\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.490877 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-config-data\") pod \"horizon-874bc5bd9-jqhms\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " pod="openstack/horizon-874bc5bd9-jqhms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.490944 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgmqf\" (UniqueName: \"kubernetes.io/projected/16339330-a7b5-471a-ad7e-1a8d4627c3e2-kube-api-access-xgmqf\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.491011 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd97l\" (UniqueName: \"kubernetes.io/projected/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-kube-api-access-bd97l\") pod \"horizon-874bc5bd9-jqhms\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " pod="openstack/horizon-874bc5bd9-jqhms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.491103 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-config\") pod \"dnsmasq-dns-785d8bcb8c-5wt4c\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.491176 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-5wt4c\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.491265 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-combined-ca-bundle\") pod \"placement-db-sync-ft76h\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " pod="openstack/placement-db-sync-ft76h" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.491336 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b349fb1-5e71-45ff-88e9-03592bfe2504-db-sync-config-data\") pod \"barbican-db-sync-sflms\" (UID: \"6b349fb1-5e71-45ff-88e9-03592bfe2504\") " pod="openstack/barbican-db-sync-sflms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.491406 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-db-sync-config-data\") pod \"cinder-db-sync-nnlq9\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.491533 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-scripts\") pod \"horizon-874bc5bd9-jqhms\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " pod="openstack/horizon-874bc5bd9-jqhms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.491609 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-5wt4c\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.491686 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-logs\") pod \"placement-db-sync-ft76h\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " pod="openstack/placement-db-sync-ft76h" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.491767 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4djz\" (UniqueName: \"kubernetes.io/projected/6b349fb1-5e71-45ff-88e9-03592bfe2504-kube-api-access-f4djz\") pod \"barbican-db-sync-sflms\" (UID: \"6b349fb1-5e71-45ff-88e9-03592bfe2504\") " pod="openstack/barbican-db-sync-sflms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.491886 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.492239 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nft6t\" (UniqueName: \"kubernetes.io/projected/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-kube-api-access-nft6t\") pod \"placement-db-sync-ft76h\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " pod="openstack/placement-db-sync-ft76h" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.492459 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-config-data\") pod \"placement-db-sync-ft76h\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " pod="openstack/placement-db-sync-ft76h" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.492562 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpwqj\" (UniqueName: \"kubernetes.io/projected/df0d5e85-642d-4fab-a77d-337c4a4518d6-kube-api-access-wpwqj\") pod \"dnsmasq-dns-785d8bcb8c-5wt4c\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.492645 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-5wt4c\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.492751 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16339330-a7b5-471a-ad7e-1a8d4627c3e2-log-httpd\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.493532 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16339330-a7b5-471a-ad7e-1a8d4627c3e2-run-httpd\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.493643 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf8c7\" (UniqueName: \"kubernetes.io/projected/c698a36a-d049-4fdb-b279-1b62d1f807d3-kube-api-access-bf8c7\") pod \"cinder-db-sync-nnlq9\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.493722 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.493812 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-horizon-secret-key\") pod \"horizon-874bc5bd9-jqhms\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " pod="openstack/horizon-874bc5bd9-jqhms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.494044 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-scripts\") pod \"placement-db-sync-ft76h\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " pod="openstack/placement-db-sync-ft76h" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.494200 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-config-data\") pod \"cinder-db-sync-nnlq9\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.494275 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-combined-ca-bundle\") pod \"cinder-db-sync-nnlq9\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.494361 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b349fb1-5e71-45ff-88e9-03592bfe2504-combined-ca-bundle\") pod \"barbican-db-sync-sflms\" (UID: \"6b349fb1-5e71-45ff-88e9-03592bfe2504\") " pod="openstack/barbican-db-sync-sflms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.494493 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-logs\") pod \"horizon-874bc5bd9-jqhms\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " pod="openstack/horizon-874bc5bd9-jqhms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.513743 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.513803 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16339330-a7b5-471a-ad7e-1a8d4627c3e2-run-httpd\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.492582 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ft76h"] Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.524191 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16339330-a7b5-471a-ad7e-1a8d4627c3e2-log-httpd\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.525704 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-config-data\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.563247 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-scripts\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.563309 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.570875 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-874bc5bd9-jqhms"] Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.594749 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgmqf\" (UniqueName: \"kubernetes.io/projected/16339330-a7b5-471a-ad7e-1a8d4627c3e2-kube-api-access-xgmqf\") pod \"ceilometer-0\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.597632 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf8c7\" (UniqueName: \"kubernetes.io/projected/c698a36a-d049-4fdb-b279-1b62d1f807d3-kube-api-access-bf8c7\") pod \"cinder-db-sync-nnlq9\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.597787 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-horizon-secret-key\") pod \"horizon-874bc5bd9-jqhms\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " pod="openstack/horizon-874bc5bd9-jqhms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.598630 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-scripts\") pod \"placement-db-sync-ft76h\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " pod="openstack/placement-db-sync-ft76h" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.598753 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-config-data\") pod \"cinder-db-sync-nnlq9\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.598852 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-combined-ca-bundle\") pod \"cinder-db-sync-nnlq9\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.598927 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b349fb1-5e71-45ff-88e9-03592bfe2504-combined-ca-bundle\") pod \"barbican-db-sync-sflms\" (UID: \"6b349fb1-5e71-45ff-88e9-03592bfe2504\") " pod="openstack/barbican-db-sync-sflms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.599014 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-logs\") pod \"horizon-874bc5bd9-jqhms\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " pod="openstack/horizon-874bc5bd9-jqhms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.599085 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c698a36a-d049-4fdb-b279-1b62d1f807d3-etc-machine-id\") pod \"cinder-db-sync-nnlq9\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.599171 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-scripts\") pod \"cinder-db-sync-nnlq9\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.599256 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-5wt4c\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.599327 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-config-data\") pod \"horizon-874bc5bd9-jqhms\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " pod="openstack/horizon-874bc5bd9-jqhms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.599500 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd97l\" (UniqueName: \"kubernetes.io/projected/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-kube-api-access-bd97l\") pod \"horizon-874bc5bd9-jqhms\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " pod="openstack/horizon-874bc5bd9-jqhms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.599596 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-5wt4c\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.599693 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-config\") pod \"dnsmasq-dns-785d8bcb8c-5wt4c\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.599770 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-combined-ca-bundle\") pod \"placement-db-sync-ft76h\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " pod="openstack/placement-db-sync-ft76h" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.599849 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b349fb1-5e71-45ff-88e9-03592bfe2504-db-sync-config-data\") pod \"barbican-db-sync-sflms\" (UID: \"6b349fb1-5e71-45ff-88e9-03592bfe2504\") " pod="openstack/barbican-db-sync-sflms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.599940 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-db-sync-config-data\") pod \"cinder-db-sync-nnlq9\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.601577 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-scripts\") pod \"horizon-874bc5bd9-jqhms\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " pod="openstack/horizon-874bc5bd9-jqhms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.601911 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-5wt4c\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.601994 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-logs\") pod \"placement-db-sync-ft76h\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " pod="openstack/placement-db-sync-ft76h" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.602120 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4djz\" (UniqueName: \"kubernetes.io/projected/6b349fb1-5e71-45ff-88e9-03592bfe2504-kube-api-access-f4djz\") pod \"barbican-db-sync-sflms\" (UID: \"6b349fb1-5e71-45ff-88e9-03592bfe2504\") " pod="openstack/barbican-db-sync-sflms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.603050 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nft6t\" (UniqueName: \"kubernetes.io/projected/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-kube-api-access-nft6t\") pod \"placement-db-sync-ft76h\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " pod="openstack/placement-db-sync-ft76h" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.603189 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-config-data\") pod \"placement-db-sync-ft76h\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " pod="openstack/placement-db-sync-ft76h" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.603276 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpwqj\" (UniqueName: \"kubernetes.io/projected/df0d5e85-642d-4fab-a77d-337c4a4518d6-kube-api-access-wpwqj\") pod \"dnsmasq-dns-785d8bcb8c-5wt4c\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.603350 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-5wt4c\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.604232 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-5wt4c\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.604522 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-config\") pod \"dnsmasq-dns-785d8bcb8c-5wt4c\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.614036 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-logs\") pod \"placement-db-sync-ft76h\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " pod="openstack/placement-db-sync-ft76h" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.618304 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-config-data\") pod \"horizon-874bc5bd9-jqhms\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " pod="openstack/horizon-874bc5bd9-jqhms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.618869 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-5wt4c\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.619026 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-logs\") pod \"horizon-874bc5bd9-jqhms\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " pod="openstack/horizon-874bc5bd9-jqhms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.619406 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c698a36a-d049-4fdb-b279-1b62d1f807d3-etc-machine-id\") pod \"cinder-db-sync-nnlq9\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.620727 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-scripts\") pod \"horizon-874bc5bd9-jqhms\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " pod="openstack/horizon-874bc5bd9-jqhms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.621594 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-5wt4c\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.622197 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-5wt4c\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.625488 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-config-data\") pod \"placement-db-sync-ft76h\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " pod="openstack/placement-db-sync-ft76h" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.628007 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b349fb1-5e71-45ff-88e9-03592bfe2504-combined-ca-bundle\") pod \"barbican-db-sync-sflms\" (UID: \"6b349fb1-5e71-45ff-88e9-03592bfe2504\") " pod="openstack/barbican-db-sync-sflms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.634741 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-horizon-secret-key\") pod \"horizon-874bc5bd9-jqhms\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " pod="openstack/horizon-874bc5bd9-jqhms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.634809 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-scripts\") pod \"placement-db-sync-ft76h\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " pod="openstack/placement-db-sync-ft76h" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.635136 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-scripts\") pod \"cinder-db-sync-nnlq9\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.635508 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-t6bn6" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.637416 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-combined-ca-bundle\") pod \"placement-db-sync-ft76h\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " pod="openstack/placement-db-sync-ft76h" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.643518 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-config-data\") pod \"cinder-db-sync-nnlq9\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.644967 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b349fb1-5e71-45ff-88e9-03592bfe2504-db-sync-config-data\") pod \"barbican-db-sync-sflms\" (UID: \"6b349fb1-5e71-45ff-88e9-03592bfe2504\") " pod="openstack/barbican-db-sync-sflms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.645974 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf8c7\" (UniqueName: \"kubernetes.io/projected/c698a36a-d049-4fdb-b279-1b62d1f807d3-kube-api-access-bf8c7\") pod \"cinder-db-sync-nnlq9\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.646028 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.648237 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.653974 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.654220 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qlfk8" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.654666 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.654802 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.670153 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.678867 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-combined-ca-bundle\") pod \"cinder-db-sync-nnlq9\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.679148 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4djz\" (UniqueName: \"kubernetes.io/projected/6b349fb1-5e71-45ff-88e9-03592bfe2504-kube-api-access-f4djz\") pod \"barbican-db-sync-sflms\" (UID: \"6b349fb1-5e71-45ff-88e9-03592bfe2504\") " pod="openstack/barbican-db-sync-sflms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.679518 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nft6t\" (UniqueName: \"kubernetes.io/projected/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-kube-api-access-nft6t\") pod \"placement-db-sync-ft76h\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " pod="openstack/placement-db-sync-ft76h" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.679722 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpwqj\" (UniqueName: \"kubernetes.io/projected/df0d5e85-642d-4fab-a77d-337c4a4518d6-kube-api-access-wpwqj\") pod \"dnsmasq-dns-785d8bcb8c-5wt4c\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.679927 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-db-sync-config-data\") pod \"cinder-db-sync-nnlq9\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.684203 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd97l\" (UniqueName: \"kubernetes.io/projected/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-kube-api-access-bd97l\") pod \"horizon-874bc5bd9-jqhms\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " pod="openstack/horizon-874bc5bd9-jqhms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.693636 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-874bc5bd9-jqhms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.711097 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ft76h" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.729260 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sflms" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.733176 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.796469 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.812814 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.813139 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-config-data\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.813204 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52998ca2-bb09-4f2f-93ce-8e759c30e420-logs\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.813238 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/52998ca2-bb09-4f2f-93ce-8e759c30e420-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.813282 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.813323 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.813350 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-scripts\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.813390 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bdz9\" (UniqueName: \"kubernetes.io/projected/52998ca2-bb09-4f2f-93ce-8e759c30e420-kube-api-access-5bdz9\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.813061 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.921751 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.921790 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-config-data\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.921875 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52998ca2-bb09-4f2f-93ce-8e759c30e420-logs\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.921918 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/52998ca2-bb09-4f2f-93ce-8e759c30e420-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.921977 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.921996 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.922026 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-scripts\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.922061 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bdz9\" (UniqueName: \"kubernetes.io/projected/52998ca2-bb09-4f2f-93ce-8e759c30e420-kube-api-access-5bdz9\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.923666 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/52998ca2-bb09-4f2f-93ce-8e759c30e420-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.923714 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52998ca2-bb09-4f2f-93ce-8e759c30e420-logs\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.924404 4904 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.944480 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.951083 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-config-data\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.952071 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.966036 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-scripts\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:29 crc kubenswrapper[4904]: I0214 11:26:29.996744 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bdz9\" (UniqueName: \"kubernetes.io/projected/52998ca2-bb09-4f2f-93ce-8e759c30e420-kube-api-access-5bdz9\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.072071 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.081977 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.083357 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.089209 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.089399 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.100336 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.127460 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.229888 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.229940 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.229961 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9564\" (UniqueName: \"kubernetes.io/projected/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-kube-api-access-j9564\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.230005 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.230041 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.230068 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.230086 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.230106 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-logs\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.250715 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" podUID="916f82bd-6f92-4e00-be47-d3343219d381" containerName="dnsmasq-dns" containerID="cri-o://ba6f63f0f7658c40d32342c79fd211328b0e6234523c5ad93b1122e4bf50d469" gracePeriod=10 Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.331853 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.331912 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.331935 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.331955 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-logs\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.332012 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.332042 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.332061 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9564\" (UniqueName: \"kubernetes.io/projected/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-kube-api-access-j9564\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.332097 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.332889 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.333227 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-logs\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.333523 4904 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.352126 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.352243 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.366717 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9564\" (UniqueName: \"kubernetes.io/projected/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-kube-api-access-j9564\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.438710 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.442603 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.449547 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.490470 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.764443 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.813072 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6b79d987b9-h8nmk"] Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.818502 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-dxj9p"] Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.829172 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-7bjcl"] Feb 14 11:26:30 crc kubenswrapper[4904]: W0214 11:26:30.882818 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0bf3b04f_1dcb_4f24_b92b_3a3e57dd46cc.slice/crio-31cf84db9c3c0de9ff6a00b9e0db66f58bf35323dfb356602f3a3881523242d6 WatchSource:0}: Error finding container 31cf84db9c3c0de9ff6a00b9e0db66f58bf35323dfb356602f3a3881523242d6: Status 404 returned error can't find the container with id 31cf84db9c3c0de9ff6a00b9e0db66f58bf35323dfb356602f3a3881523242d6 Feb 14 11:26:30 crc kubenswrapper[4904]: I0214 11:26:30.982598 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-t6bn6"] Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.146855 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-nnlq9"] Feb 14 11:26:31 crc kubenswrapper[4904]: W0214 11:26:31.154170 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc698a36a_d049_4fdb_b279_1b62d1f807d3.slice/crio-ed32c9db7ff9f74d2e4fdc9f70965da4f5969211cd1e7f3d9a5e4641d4515841 WatchSource:0}: Error finding container ed32c9db7ff9f74d2e4fdc9f70965da4f5969211cd1e7f3d9a5e4641d4515841: Status 404 returned error can't find the container with id ed32c9db7ff9f74d2e4fdc9f70965da4f5969211cd1e7f3d9a5e4641d4515841 Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.284111 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b79d987b9-h8nmk" event={"ID":"97bc1e53-5edf-43fc-8aca-0dee904c86eb","Type":"ContainerStarted","Data":"8d966d7fb7d3fd39ec5fa4d4980bb6d767da93884c37a8955b66bb3bbc7130c0"} Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.287259 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7bjcl" event={"ID":"51d73444-44fa-47bf-b8a4-d5c46b0cf80a","Type":"ContainerStarted","Data":"59c338cdaf7c0b4c5dcee4c5efb50931c86534e944cfcedccaad2601ebf22c65"} Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.299460 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-t6bn6" event={"ID":"2abcf43f-6f3b-4804-83a5-ba42fc66f2ec","Type":"ContainerStarted","Data":"24bf3019ed6c686bcb675250b277d211fef8c3300ed240bac657ff502d141bda"} Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.310076 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" event={"ID":"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc","Type":"ContainerStarted","Data":"31cf84db9c3c0de9ff6a00b9e0db66f58bf35323dfb356602f3a3881523242d6"} Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.347802 4904 generic.go:334] "Generic (PLEG): container finished" podID="916f82bd-6f92-4e00-be47-d3343219d381" containerID="ba6f63f0f7658c40d32342c79fd211328b0e6234523c5ad93b1122e4bf50d469" exitCode=0 Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.347888 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" event={"ID":"916f82bd-6f92-4e00-be47-d3343219d381","Type":"ContainerDied","Data":"ba6f63f0f7658c40d32342c79fd211328b0e6234523c5ad93b1122e4bf50d469"} Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.349146 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nnlq9" event={"ID":"c698a36a-d049-4fdb-b279-1b62d1f807d3","Type":"ContainerStarted","Data":"ed32c9db7ff9f74d2e4fdc9f70965da4f5969211cd1e7f3d9a5e4641d4515841"} Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.530582 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-874bc5bd9-jqhms"] Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.558652 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ft76h"] Feb 14 11:26:31 crc kubenswrapper[4904]: W0214 11:26:31.603950 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60ac5de4_7c8e_4065_a846_3de7a3cffcc0.slice/crio-541a3dfa767b27a430b04c055cfc2580684232830d887cddbc03fc91ba12fdc9 WatchSource:0}: Error finding container 541a3dfa767b27a430b04c055cfc2580684232830d887cddbc03fc91ba12fdc9: Status 404 returned error can't find the container with id 541a3dfa767b27a430b04c055cfc2580684232830d887cddbc03fc91ba12fdc9 Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.646943 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-sflms"] Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.785249 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:26:31 crc kubenswrapper[4904]: W0214 11:26:31.801444 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16339330_a7b5_471a_ad7e_1a8d4627c3e2.slice/crio-deb50cca9f879e4b9ec57f6beab522df24ed4ba1a8d60a33e487fe0676797254 WatchSource:0}: Error finding container deb50cca9f879e4b9ec57f6beab522df24ed4ba1a8d60a33e487fe0676797254: Status 404 returned error can't find the container with id deb50cca9f879e4b9ec57f6beab522df24ed4ba1a8d60a33e487fe0676797254 Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.829612 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-5wt4c"] Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.866772 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.977798 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-config\") pod \"916f82bd-6f92-4e00-be47-d3343219d381\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.978385 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-ovsdbserver-nb\") pod \"916f82bd-6f92-4e00-be47-d3343219d381\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.978458 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hntdn\" (UniqueName: \"kubernetes.io/projected/916f82bd-6f92-4e00-be47-d3343219d381-kube-api-access-hntdn\") pod \"916f82bd-6f92-4e00-be47-d3343219d381\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.978490 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-ovsdbserver-sb\") pod \"916f82bd-6f92-4e00-be47-d3343219d381\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.978580 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-dns-svc\") pod \"916f82bd-6f92-4e00-be47-d3343219d381\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " Feb 14 11:26:31 crc kubenswrapper[4904]: I0214 11:26:31.978650 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-dns-swift-storage-0\") pod \"916f82bd-6f92-4e00-be47-d3343219d381\" (UID: \"916f82bd-6f92-4e00-be47-d3343219d381\") " Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.006906 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/916f82bd-6f92-4e00-be47-d3343219d381-kube-api-access-hntdn" (OuterVolumeSpecName: "kube-api-access-hntdn") pod "916f82bd-6f92-4e00-be47-d3343219d381" (UID: "916f82bd-6f92-4e00-be47-d3343219d381"). InnerVolumeSpecName "kube-api-access-hntdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.045667 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.063392 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "916f82bd-6f92-4e00-be47-d3343219d381" (UID: "916f82bd-6f92-4e00-be47-d3343219d381"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.079641 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "916f82bd-6f92-4e00-be47-d3343219d381" (UID: "916f82bd-6f92-4e00-be47-d3343219d381"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.081168 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.081189 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hntdn\" (UniqueName: \"kubernetes.io/projected/916f82bd-6f92-4e00-be47-d3343219d381-kube-api-access-hntdn\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.081200 4904 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.121056 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "916f82bd-6f92-4e00-be47-d3343219d381" (UID: "916f82bd-6f92-4e00-be47-d3343219d381"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.167495 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "916f82bd-6f92-4e00-be47-d3343219d381" (UID: "916f82bd-6f92-4e00-be47-d3343219d381"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.168505 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-config" (OuterVolumeSpecName: "config") pod "916f82bd-6f92-4e00-be47-d3343219d381" (UID: "916f82bd-6f92-4e00-be47-d3343219d381"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.195177 4904 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.195250 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.195328 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/916f82bd-6f92-4e00-be47-d3343219d381-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.402151 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7","Type":"ContainerStarted","Data":"0ec617f10d59e740c31d03aa2d9da1223ed2160f8a32c7671e32c9dc1ceef7c5"} Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.445234 4904 generic.go:334] "Generic (PLEG): container finished" podID="0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc" containerID="bd2341627c8a5d70aaadd4c711d3d91ce908b961eedc4321780a20b94ab1192e" exitCode=0 Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.445340 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" event={"ID":"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc","Type":"ContainerDied","Data":"bd2341627c8a5d70aaadd4c711d3d91ce908b961eedc4321780a20b94ab1192e"} Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.455245 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" event={"ID":"916f82bd-6f92-4e00-be47-d3343219d381","Type":"ContainerDied","Data":"3cbc0cee9bb5d92e61e8983e7bcb2a6485fc88156a858f8bc209e98b9d39d58a"} Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.455293 4904 scope.go:117] "RemoveContainer" containerID="ba6f63f0f7658c40d32342c79fd211328b0e6234523c5ad93b1122e4bf50d469" Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.455398 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-lrs72" Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.462721 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16339330-a7b5-471a-ad7e-1a8d4627c3e2","Type":"ContainerStarted","Data":"deb50cca9f879e4b9ec57f6beab522df24ed4ba1a8d60a33e487fe0676797254"} Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.476004 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" event={"ID":"df0d5e85-642d-4fab-a77d-337c4a4518d6","Type":"ContainerStarted","Data":"39b62eecc442c6db7887ed31c8ff69711b14ec1aaf3b11345182a53fe3ce70df"} Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.480277 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sflms" event={"ID":"6b349fb1-5e71-45ff-88e9-03592bfe2504","Type":"ContainerStarted","Data":"903e1842b32d1188981baed3b1e4cd5f7449328af56439f5dc20700bd5617a34"} Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.509132 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-lrs72"] Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.513981 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7bjcl" event={"ID":"51d73444-44fa-47bf-b8a4-d5c46b0cf80a","Type":"ContainerStarted","Data":"6bb8ce5e37d9faa7fb04dbc186646c0baff6a994d8e2887ed71781d05f2a5b2d"} Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.517883 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-874bc5bd9-jqhms" event={"ID":"0c6ef645-08d7-4bd0-aa8a-8181bae77b28","Type":"ContainerStarted","Data":"74057c0f494e7b7cf7b23959a450f5cc749a391519fca50368d0a9113e2d7193"} Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.524199 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-lrs72"] Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.525917 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-t6bn6" event={"ID":"2abcf43f-6f3b-4804-83a5-ba42fc66f2ec","Type":"ContainerStarted","Data":"41db167ff7dc3e24f9b2ac1e32c2405e16f5f00ee2611e44e5400ce8318cbd67"} Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.533006 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ft76h" event={"ID":"60ac5de4-7c8e-4065-a846-3de7a3cffcc0","Type":"ContainerStarted","Data":"541a3dfa767b27a430b04c055cfc2580684232830d887cddbc03fc91ba12fdc9"} Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.548130 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-7bjcl" podStartSLOduration=4.54811011 podStartE2EDuration="4.54811011s" podCreationTimestamp="2026-02-14 11:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:26:32.542182728 +0000 UTC m=+983.354947389" watchObservedRunningTime="2026-02-14 11:26:32.54811011 +0000 UTC m=+983.360874771" Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.568995 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-t6bn6" podStartSLOduration=4.568976963 podStartE2EDuration="4.568976963s" podCreationTimestamp="2026-02-14 11:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:26:32.566949827 +0000 UTC m=+983.379714488" watchObservedRunningTime="2026-02-14 11:26:32.568976963 +0000 UTC m=+983.381741624" Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.571685 4904 scope.go:117] "RemoveContainer" containerID="ef1ad53d02639032e085934ce28fe8a2f0f54d525265966d6d22e0c72e12a9c7" Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.918219 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nbqrf" Feb 14 11:26:32 crc kubenswrapper[4904]: I0214 11:26:32.924606 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nbqrf" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.065921 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nbqrf" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.070353 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6b79d987b9-h8nmk"] Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.093506 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.123922 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.195324 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.222041 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8v2xr" podUID="a9fd9edc-0058-417c-baa6-9fbff5035af5" containerName="registry-server" probeResult="failure" output=< Feb 14 11:26:33 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 11:26:33 crc kubenswrapper[4904]: > Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.265678 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-569f4ff6dc-kvsmj"] Feb 14 11:26:33 crc kubenswrapper[4904]: E0214 11:26:33.277165 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc" containerName="init" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.278799 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc" containerName="init" Feb 14 11:26:33 crc kubenswrapper[4904]: E0214 11:26:33.278933 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="916f82bd-6f92-4e00-be47-d3343219d381" containerName="init" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.278991 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="916f82bd-6f92-4e00-be47-d3343219d381" containerName="init" Feb 14 11:26:33 crc kubenswrapper[4904]: E0214 11:26:33.279094 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="916f82bd-6f92-4e00-be47-d3343219d381" containerName="dnsmasq-dns" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.279151 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="916f82bd-6f92-4e00-be47-d3343219d381" containerName="dnsmasq-dns" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.279887 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc" containerName="init" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.279975 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="916f82bd-6f92-4e00-be47-d3343219d381" containerName="dnsmasq-dns" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.282681 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-569f4ff6dc-kvsmj" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.340908 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-dns-swift-storage-0\") pod \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.340969 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-dns-svc\") pod \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.340997 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-ovsdbserver-sb\") pod \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.341056 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpmc2\" (UniqueName: \"kubernetes.io/projected/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-kube-api-access-tpmc2\") pod \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.341110 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-ovsdbserver-nb\") pod \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.341219 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-config\") pod \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.341686 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5q25\" (UniqueName: \"kubernetes.io/projected/fa07aa03-b19b-42ba-9834-0ce67c1df067-kube-api-access-v5q25\") pod \"horizon-569f4ff6dc-kvsmj\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " pod="openstack/horizon-569f4ff6dc-kvsmj" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.341741 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fa07aa03-b19b-42ba-9834-0ce67c1df067-config-data\") pod \"horizon-569f4ff6dc-kvsmj\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " pod="openstack/horizon-569f4ff6dc-kvsmj" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.341763 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa07aa03-b19b-42ba-9834-0ce67c1df067-scripts\") pod \"horizon-569f4ff6dc-kvsmj\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " pod="openstack/horizon-569f4ff6dc-kvsmj" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.341792 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa07aa03-b19b-42ba-9834-0ce67c1df067-logs\") pod \"horizon-569f4ff6dc-kvsmj\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " pod="openstack/horizon-569f4ff6dc-kvsmj" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.341813 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fa07aa03-b19b-42ba-9834-0ce67c1df067-horizon-secret-key\") pod \"horizon-569f4ff6dc-kvsmj\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " pod="openstack/horizon-569f4ff6dc-kvsmj" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.457127 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-config" (OuterVolumeSpecName: "config") pod "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc" (UID: "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.491249 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc" (UID: "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.492590 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc" (UID: "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.492722 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-kube-api-access-tpmc2" (OuterVolumeSpecName: "kube-api-access-tpmc2") pod "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc" (UID: "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc"). InnerVolumeSpecName "kube-api-access-tpmc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.493842 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-config\") pod \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.493941 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-dns-swift-storage-0\") pod \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.493990 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-ovsdbserver-sb\") pod \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.494054 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpmc2\" (UniqueName: \"kubernetes.io/projected/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-kube-api-access-tpmc2\") pod \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\" (UID: \"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc\") " Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.494377 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fa07aa03-b19b-42ba-9834-0ce67c1df067-config-data\") pod \"horizon-569f4ff6dc-kvsmj\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " pod="openstack/horizon-569f4ff6dc-kvsmj" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.494409 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa07aa03-b19b-42ba-9834-0ce67c1df067-scripts\") pod \"horizon-569f4ff6dc-kvsmj\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " pod="openstack/horizon-569f4ff6dc-kvsmj" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.494445 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa07aa03-b19b-42ba-9834-0ce67c1df067-logs\") pod \"horizon-569f4ff6dc-kvsmj\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " pod="openstack/horizon-569f4ff6dc-kvsmj" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.494478 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fa07aa03-b19b-42ba-9834-0ce67c1df067-horizon-secret-key\") pod \"horizon-569f4ff6dc-kvsmj\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " pod="openstack/horizon-569f4ff6dc-kvsmj" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.494688 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5q25\" (UniqueName: \"kubernetes.io/projected/fa07aa03-b19b-42ba-9834-0ce67c1df067-kube-api-access-v5q25\") pod \"horizon-569f4ff6dc-kvsmj\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " pod="openstack/horizon-569f4ff6dc-kvsmj" Feb 14 11:26:33 crc kubenswrapper[4904]: W0214 11:26:33.496134 4904 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc/volumes/kubernetes.io~configmap/config Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.496146 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-config" (OuterVolumeSpecName: "config") pod "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc" (UID: "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:33 crc kubenswrapper[4904]: W0214 11:26:33.496185 4904 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc/volumes/kubernetes.io~configmap/dns-swift-storage-0 Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.496191 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc" (UID: "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:33 crc kubenswrapper[4904]: W0214 11:26:33.496224 4904 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc/volumes/kubernetes.io~configmap/ovsdbserver-sb Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.496230 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc" (UID: "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:33 crc kubenswrapper[4904]: W0214 11:26:33.496264 4904 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc/volumes/kubernetes.io~projected/kube-api-access-tpmc2 Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.496300 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-kube-api-access-tpmc2" (OuterVolumeSpecName: "kube-api-access-tpmc2") pod "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc" (UID: "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc"). InnerVolumeSpecName "kube-api-access-tpmc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.498977 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-569f4ff6dc-kvsmj"] Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.499808 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa07aa03-b19b-42ba-9834-0ce67c1df067-logs\") pod \"horizon-569f4ff6dc-kvsmj\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " pod="openstack/horizon-569f4ff6dc-kvsmj" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.500541 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa07aa03-b19b-42ba-9834-0ce67c1df067-scripts\") pod \"horizon-569f4ff6dc-kvsmj\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " pod="openstack/horizon-569f4ff6dc-kvsmj" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.516787 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fa07aa03-b19b-42ba-9834-0ce67c1df067-config-data\") pod \"horizon-569f4ff6dc-kvsmj\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " pod="openstack/horizon-569f4ff6dc-kvsmj" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.516868 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.548396 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5q25\" (UniqueName: \"kubernetes.io/projected/fa07aa03-b19b-42ba-9834-0ce67c1df067-kube-api-access-v5q25\") pod \"horizon-569f4ff6dc-kvsmj\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " pod="openstack/horizon-569f4ff6dc-kvsmj" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.556554 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.558684 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc" (UID: "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.567246 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc" (UID: "0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.570575 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fa07aa03-b19b-42ba-9834-0ce67c1df067-horizon-secret-key\") pod \"horizon-569f4ff6dc-kvsmj\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " pod="openstack/horizon-569f4ff6dc-kvsmj" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.598244 4904 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.598538 4904 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.598548 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.598557 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpmc2\" (UniqueName: \"kubernetes.io/projected/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-kube-api-access-tpmc2\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.598566 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.598574 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.619066 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-569f4ff6dc-kvsmj" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.667248 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" event={"ID":"0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc","Type":"ContainerDied","Data":"31cf84db9c3c0de9ff6a00b9e0db66f58bf35323dfb356602f3a3881523242d6"} Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.667296 4904 scope.go:117] "RemoveContainer" containerID="bd2341627c8a5d70aaadd4c711d3d91ce908b961eedc4321780a20b94ab1192e" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.667415 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-dxj9p" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.754383 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-dxj9p"] Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.762280 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"52998ca2-bb09-4f2f-93ce-8e759c30e420","Type":"ContainerStarted","Data":"23939baaf6bb40781b668f728fdc7b5782eafefcfcbde735eff1d4476279976f"} Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.763772 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-dxj9p"] Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.774794 4904 generic.go:334] "Generic (PLEG): container finished" podID="df0d5e85-642d-4fab-a77d-337c4a4518d6" containerID="46d5f87b0312e65ebe0479ad137dabccf046ed0c414e4b6ba23423036fcc0147" exitCode=0 Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.776743 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" event={"ID":"df0d5e85-642d-4fab-a77d-337c4a4518d6","Type":"ContainerDied","Data":"46d5f87b0312e65ebe0479ad137dabccf046ed0c414e4b6ba23423036fcc0147"} Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.870621 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc" path="/var/lib/kubelet/pods/0bf3b04f-1dcb-4f24-b92b-3a3e57dd46cc/volumes" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.872033 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="916f82bd-6f92-4e00-be47-d3343219d381" path="/var/lib/kubelet/pods/916f82bd-6f92-4e00-be47-d3343219d381/volumes" Feb 14 11:26:33 crc kubenswrapper[4904]: I0214 11:26:33.918678 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nbqrf" Feb 14 11:26:34 crc kubenswrapper[4904]: I0214 11:26:34.006057 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbqrf"] Feb 14 11:26:34 crc kubenswrapper[4904]: I0214 11:26:34.251654 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-569f4ff6dc-kvsmj"] Feb 14 11:26:34 crc kubenswrapper[4904]: I0214 11:26:34.788431 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7","Type":"ContainerStarted","Data":"ab52531d979b58f2b6af5732827f39d98c54f479d4f0eb2695e50e4284f0359e"} Feb 14 11:26:34 crc kubenswrapper[4904]: I0214 11:26:34.807410 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-569f4ff6dc-kvsmj" event={"ID":"fa07aa03-b19b-42ba-9834-0ce67c1df067","Type":"ContainerStarted","Data":"5b326289aa04248f581504c47df3509463c6e29f3bbd7ed3ffc4925d788fb974"} Feb 14 11:26:34 crc kubenswrapper[4904]: I0214 11:26:34.817598 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" event={"ID":"df0d5e85-642d-4fab-a77d-337c4a4518d6","Type":"ContainerStarted","Data":"613f7e92e44f4c81d170f628b5fcb673319c5698a425bc631d940c9ecbff4c13"} Feb 14 11:26:34 crc kubenswrapper[4904]: I0214 11:26:34.817647 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:34 crc kubenswrapper[4904]: I0214 11:26:34.847093 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" podStartSLOduration=5.847074434 podStartE2EDuration="5.847074434s" podCreationTimestamp="2026-02-14 11:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:26:34.844226036 +0000 UTC m=+985.656990697" watchObservedRunningTime="2026-02-14 11:26:34.847074434 +0000 UTC m=+985.659839085" Feb 14 11:26:35 crc kubenswrapper[4904]: I0214 11:26:35.863129 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nbqrf" podUID="fd725c16-a4a1-4dca-b7dc-a159d272f4b0" containerName="registry-server" containerID="cri-o://58f134133db6d54efd6509fd24fa849311a43dc0c4a199b39e5681da001f7985" gracePeriod=2 Feb 14 11:26:35 crc kubenswrapper[4904]: I0214 11:26:35.865513 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7" containerName="glance-log" containerID="cri-o://ab52531d979b58f2b6af5732827f39d98c54f479d4f0eb2695e50e4284f0359e" gracePeriod=30 Feb 14 11:26:35 crc kubenswrapper[4904]: I0214 11:26:35.865676 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7" containerName="glance-httpd" containerID="cri-o://6b8569215b127848980d65bb56dbb8592e5097497d0e15a1b3593b952c0764d3" gracePeriod=30 Feb 14 11:26:35 crc kubenswrapper[4904]: I0214 11:26:35.880712 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"52998ca2-bb09-4f2f-93ce-8e759c30e420","Type":"ContainerStarted","Data":"dab00463f4f9f87f8e5fd9430d330e67d0b76a0c859df4d3abe75d67fbcfaf80"} Feb 14 11:26:35 crc kubenswrapper[4904]: I0214 11:26:35.964413 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.964395328 podStartE2EDuration="7.964395328s" podCreationTimestamp="2026-02-14 11:26:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:26:35.946507628 +0000 UTC m=+986.759272289" watchObservedRunningTime="2026-02-14 11:26:35.964395328 +0000 UTC m=+986.777159990" Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.826622 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nbqrf" Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.884280 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6wkq\" (UniqueName: \"kubernetes.io/projected/fd725c16-a4a1-4dca-b7dc-a159d272f4b0-kube-api-access-w6wkq\") pod \"fd725c16-a4a1-4dca-b7dc-a159d272f4b0\" (UID: \"fd725c16-a4a1-4dca-b7dc-a159d272f4b0\") " Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.884374 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd725c16-a4a1-4dca-b7dc-a159d272f4b0-utilities\") pod \"fd725c16-a4a1-4dca-b7dc-a159d272f4b0\" (UID: \"fd725c16-a4a1-4dca-b7dc-a159d272f4b0\") " Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.884436 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd725c16-a4a1-4dca-b7dc-a159d272f4b0-catalog-content\") pod \"fd725c16-a4a1-4dca-b7dc-a159d272f4b0\" (UID: \"fd725c16-a4a1-4dca-b7dc-a159d272f4b0\") " Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.887975 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd725c16-a4a1-4dca-b7dc-a159d272f4b0-utilities" (OuterVolumeSpecName: "utilities") pod "fd725c16-a4a1-4dca-b7dc-a159d272f4b0" (UID: "fd725c16-a4a1-4dca-b7dc-a159d272f4b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.890053 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"52998ca2-bb09-4f2f-93ce-8e759c30e420","Type":"ContainerStarted","Data":"ab57ab2fb1dceb375552d44e51e81505405e2d336ed8cb4e17c1612f7bfadba4"} Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.892946 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd725c16-a4a1-4dca-b7dc-a159d272f4b0-kube-api-access-w6wkq" (OuterVolumeSpecName: "kube-api-access-w6wkq") pod "fd725c16-a4a1-4dca-b7dc-a159d272f4b0" (UID: "fd725c16-a4a1-4dca-b7dc-a159d272f4b0"). InnerVolumeSpecName "kube-api-access-w6wkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.894171 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="52998ca2-bb09-4f2f-93ce-8e759c30e420" containerName="glance-log" containerID="cri-o://dab00463f4f9f87f8e5fd9430d330e67d0b76a0c859df4d3abe75d67fbcfaf80" gracePeriod=30 Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.894773 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="52998ca2-bb09-4f2f-93ce-8e759c30e420" containerName="glance-httpd" containerID="cri-o://ab57ab2fb1dceb375552d44e51e81505405e2d336ed8cb4e17c1612f7bfadba4" gracePeriod=30 Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.917667 4904 generic.go:334] "Generic (PLEG): container finished" podID="f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7" containerID="6b8569215b127848980d65bb56dbb8592e5097497d0e15a1b3593b952c0764d3" exitCode=143 Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.917892 4904 generic.go:334] "Generic (PLEG): container finished" podID="f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7" containerID="ab52531d979b58f2b6af5732827f39d98c54f479d4f0eb2695e50e4284f0359e" exitCode=143 Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.917743 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7","Type":"ContainerDied","Data":"6b8569215b127848980d65bb56dbb8592e5097497d0e15a1b3593b952c0764d3"} Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.918098 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7","Type":"ContainerDied","Data":"ab52531d979b58f2b6af5732827f39d98c54f479d4f0eb2695e50e4284f0359e"} Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.952887 4904 generic.go:334] "Generic (PLEG): container finished" podID="fd725c16-a4a1-4dca-b7dc-a159d272f4b0" containerID="58f134133db6d54efd6509fd24fa849311a43dc0c4a199b39e5681da001f7985" exitCode=0 Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.952964 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbqrf" event={"ID":"fd725c16-a4a1-4dca-b7dc-a159d272f4b0","Type":"ContainerDied","Data":"58f134133db6d54efd6509fd24fa849311a43dc0c4a199b39e5681da001f7985"} Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.952993 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbqrf" event={"ID":"fd725c16-a4a1-4dca-b7dc-a159d272f4b0","Type":"ContainerDied","Data":"9c03e5b377a38f43a5317a99e4ddb964e518e6f2e15dbf8f7c0e1e1d37fbd5ae"} Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.953011 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nbqrf" Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.953034 4904 scope.go:117] "RemoveContainer" containerID="58f134133db6d54efd6509fd24fa849311a43dc0c4a199b39e5681da001f7985" Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.959806 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.9597694 podStartE2EDuration="7.9597694s" podCreationTimestamp="2026-02-14 11:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:26:36.939575797 +0000 UTC m=+987.752340458" watchObservedRunningTime="2026-02-14 11:26:36.9597694 +0000 UTC m=+987.772534051" Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.968516 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd725c16-a4a1-4dca-b7dc-a159d272f4b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fd725c16-a4a1-4dca-b7dc-a159d272f4b0" (UID: "fd725c16-a4a1-4dca-b7dc-a159d272f4b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.986822 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6wkq\" (UniqueName: \"kubernetes.io/projected/fd725c16-a4a1-4dca-b7dc-a159d272f4b0-kube-api-access-w6wkq\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.986870 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd725c16-a4a1-4dca-b7dc-a159d272f4b0-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:36 crc kubenswrapper[4904]: I0214 11:26:36.986879 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd725c16-a4a1-4dca-b7dc-a159d272f4b0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.069524 4904 scope.go:117] "RemoveContainer" containerID="0d444ca3e2b3c8d4fd3da4f79b5df977d653d191a3fff863c754d792a27f0d5e" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.183280 4904 scope.go:117] "RemoveContainer" containerID="200a3f4b00beaba2c6572617b88e7d9657462037d59c847932ce70f8e1b693c3" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.244708 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.266469 4904 scope.go:117] "RemoveContainer" containerID="58f134133db6d54efd6509fd24fa849311a43dc0c4a199b39e5681da001f7985" Feb 14 11:26:37 crc kubenswrapper[4904]: E0214 11:26:37.270587 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58f134133db6d54efd6509fd24fa849311a43dc0c4a199b39e5681da001f7985\": container with ID starting with 58f134133db6d54efd6509fd24fa849311a43dc0c4a199b39e5681da001f7985 not found: ID does not exist" containerID="58f134133db6d54efd6509fd24fa849311a43dc0c4a199b39e5681da001f7985" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.270625 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58f134133db6d54efd6509fd24fa849311a43dc0c4a199b39e5681da001f7985"} err="failed to get container status \"58f134133db6d54efd6509fd24fa849311a43dc0c4a199b39e5681da001f7985\": rpc error: code = NotFound desc = could not find container \"58f134133db6d54efd6509fd24fa849311a43dc0c4a199b39e5681da001f7985\": container with ID starting with 58f134133db6d54efd6509fd24fa849311a43dc0c4a199b39e5681da001f7985 not found: ID does not exist" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.270645 4904 scope.go:117] "RemoveContainer" containerID="0d444ca3e2b3c8d4fd3da4f79b5df977d653d191a3fff863c754d792a27f0d5e" Feb 14 11:26:37 crc kubenswrapper[4904]: E0214 11:26:37.274483 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d444ca3e2b3c8d4fd3da4f79b5df977d653d191a3fff863c754d792a27f0d5e\": container with ID starting with 0d444ca3e2b3c8d4fd3da4f79b5df977d653d191a3fff863c754d792a27f0d5e not found: ID does not exist" containerID="0d444ca3e2b3c8d4fd3da4f79b5df977d653d191a3fff863c754d792a27f0d5e" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.274521 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d444ca3e2b3c8d4fd3da4f79b5df977d653d191a3fff863c754d792a27f0d5e"} err="failed to get container status \"0d444ca3e2b3c8d4fd3da4f79b5df977d653d191a3fff863c754d792a27f0d5e\": rpc error: code = NotFound desc = could not find container \"0d444ca3e2b3c8d4fd3da4f79b5df977d653d191a3fff863c754d792a27f0d5e\": container with ID starting with 0d444ca3e2b3c8d4fd3da4f79b5df977d653d191a3fff863c754d792a27f0d5e not found: ID does not exist" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.274544 4904 scope.go:117] "RemoveContainer" containerID="200a3f4b00beaba2c6572617b88e7d9657462037d59c847932ce70f8e1b693c3" Feb 14 11:26:37 crc kubenswrapper[4904]: E0214 11:26:37.274963 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"200a3f4b00beaba2c6572617b88e7d9657462037d59c847932ce70f8e1b693c3\": container with ID starting with 200a3f4b00beaba2c6572617b88e7d9657462037d59c847932ce70f8e1b693c3 not found: ID does not exist" containerID="200a3f4b00beaba2c6572617b88e7d9657462037d59c847932ce70f8e1b693c3" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.274986 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"200a3f4b00beaba2c6572617b88e7d9657462037d59c847932ce70f8e1b693c3"} err="failed to get container status \"200a3f4b00beaba2c6572617b88e7d9657462037d59c847932ce70f8e1b693c3\": rpc error: code = NotFound desc = could not find container \"200a3f4b00beaba2c6572617b88e7d9657462037d59c847932ce70f8e1b693c3\": container with ID starting with 200a3f4b00beaba2c6572617b88e7d9657462037d59c847932ce70f8e1b693c3 not found: ID does not exist" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.293463 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-scripts\") pod \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.293542 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-httpd-run\") pod \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.293608 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-logs\") pod \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.293631 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-internal-tls-certs\") pod \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.293669 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.293702 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9564\" (UniqueName: \"kubernetes.io/projected/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-kube-api-access-j9564\") pod \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.293725 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-combined-ca-bundle\") pod \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.293765 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-config-data\") pod \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\" (UID: \"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7\") " Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.308390 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbqrf"] Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.309179 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7" (UID: "f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.309388 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-logs" (OuterVolumeSpecName: "logs") pod "f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7" (UID: "f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.318164 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbqrf"] Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.327434 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-kube-api-access-j9564" (OuterVolumeSpecName: "kube-api-access-j9564") pod "f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7" (UID: "f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7"). InnerVolumeSpecName "kube-api-access-j9564". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.334684 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7" (UID: "f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.337225 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-scripts" (OuterVolumeSpecName: "scripts") pod "f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7" (UID: "f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.363126 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7" (UID: "f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.395432 4904 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.395465 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9564\" (UniqueName: \"kubernetes.io/projected/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-kube-api-access-j9564\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.395497 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.395507 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.395516 4904 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.395524 4904 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-logs\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.398294 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-config-data" (OuterVolumeSpecName: "config-data") pod "f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7" (UID: "f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.406729 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7" (UID: "f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.429258 4904 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.497786 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.497826 4904 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.497852 4904 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.704598 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.804354 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52998ca2-bb09-4f2f-93ce-8e759c30e420-logs\") pod \"52998ca2-bb09-4f2f-93ce-8e759c30e420\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.804395 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-combined-ca-bundle\") pod \"52998ca2-bb09-4f2f-93ce-8e759c30e420\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.804562 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-config-data\") pod \"52998ca2-bb09-4f2f-93ce-8e759c30e420\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.804602 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bdz9\" (UniqueName: \"kubernetes.io/projected/52998ca2-bb09-4f2f-93ce-8e759c30e420-kube-api-access-5bdz9\") pod \"52998ca2-bb09-4f2f-93ce-8e759c30e420\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.804648 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-scripts\") pod \"52998ca2-bb09-4f2f-93ce-8e759c30e420\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.804672 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-public-tls-certs\") pod \"52998ca2-bb09-4f2f-93ce-8e759c30e420\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.804693 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/52998ca2-bb09-4f2f-93ce-8e759c30e420-httpd-run\") pod \"52998ca2-bb09-4f2f-93ce-8e759c30e420\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.804759 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"52998ca2-bb09-4f2f-93ce-8e759c30e420\" (UID: \"52998ca2-bb09-4f2f-93ce-8e759c30e420\") " Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.804900 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52998ca2-bb09-4f2f-93ce-8e759c30e420-logs" (OuterVolumeSpecName: "logs") pod "52998ca2-bb09-4f2f-93ce-8e759c30e420" (UID: "52998ca2-bb09-4f2f-93ce-8e759c30e420"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.805333 4904 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52998ca2-bb09-4f2f-93ce-8e759c30e420-logs\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.805456 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52998ca2-bb09-4f2f-93ce-8e759c30e420-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "52998ca2-bb09-4f2f-93ce-8e759c30e420" (UID: "52998ca2-bb09-4f2f-93ce-8e759c30e420"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.814189 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "52998ca2-bb09-4f2f-93ce-8e759c30e420" (UID: "52998ca2-bb09-4f2f-93ce-8e759c30e420"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.815017 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52998ca2-bb09-4f2f-93ce-8e759c30e420-kube-api-access-5bdz9" (OuterVolumeSpecName: "kube-api-access-5bdz9") pod "52998ca2-bb09-4f2f-93ce-8e759c30e420" (UID: "52998ca2-bb09-4f2f-93ce-8e759c30e420"). InnerVolumeSpecName "kube-api-access-5bdz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.818770 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-scripts" (OuterVolumeSpecName: "scripts") pod "52998ca2-bb09-4f2f-93ce-8e759c30e420" (UID: "52998ca2-bb09-4f2f-93ce-8e759c30e420"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.853959 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd725c16-a4a1-4dca-b7dc-a159d272f4b0" path="/var/lib/kubelet/pods/fd725c16-a4a1-4dca-b7dc-a159d272f4b0/volumes" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.857752 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "52998ca2-bb09-4f2f-93ce-8e759c30e420" (UID: "52998ca2-bb09-4f2f-93ce-8e759c30e420"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.894581 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-config-data" (OuterVolumeSpecName: "config-data") pod "52998ca2-bb09-4f2f-93ce-8e759c30e420" (UID: "52998ca2-bb09-4f2f-93ce-8e759c30e420"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.900721 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "52998ca2-bb09-4f2f-93ce-8e759c30e420" (UID: "52998ca2-bb09-4f2f-93ce-8e759c30e420"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.906795 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.906828 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bdz9\" (UniqueName: \"kubernetes.io/projected/52998ca2-bb09-4f2f-93ce-8e759c30e420-kube-api-access-5bdz9\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.906858 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.906867 4904 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.906878 4904 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/52998ca2-bb09-4f2f-93ce-8e759c30e420-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.906912 4904 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.906920 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52998ca2-bb09-4f2f-93ce-8e759c30e420-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.944281 4904 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.996147 4904 generic.go:334] "Generic (PLEG): container finished" podID="52998ca2-bb09-4f2f-93ce-8e759c30e420" containerID="ab57ab2fb1dceb375552d44e51e81505405e2d336ed8cb4e17c1612f7bfadba4" exitCode=143 Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.996180 4904 generic.go:334] "Generic (PLEG): container finished" podID="52998ca2-bb09-4f2f-93ce-8e759c30e420" containerID="dab00463f4f9f87f8e5fd9430d330e67d0b76a0c859df4d3abe75d67fbcfaf80" exitCode=143 Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.996259 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"52998ca2-bb09-4f2f-93ce-8e759c30e420","Type":"ContainerDied","Data":"ab57ab2fb1dceb375552d44e51e81505405e2d336ed8cb4e17c1612f7bfadba4"} Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.996290 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"52998ca2-bb09-4f2f-93ce-8e759c30e420","Type":"ContainerDied","Data":"dab00463f4f9f87f8e5fd9430d330e67d0b76a0c859df4d3abe75d67fbcfaf80"} Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.996262 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.996301 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"52998ca2-bb09-4f2f-93ce-8e759c30e420","Type":"ContainerDied","Data":"23939baaf6bb40781b668f728fdc7b5782eafefcfcbde735eff1d4476279976f"} Feb 14 11:26:37 crc kubenswrapper[4904]: I0214 11:26:37.996319 4904 scope.go:117] "RemoveContainer" containerID="ab57ab2fb1dceb375552d44e51e81505405e2d336ed8cb4e17c1612f7bfadba4" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.004351 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7","Type":"ContainerDied","Data":"0ec617f10d59e740c31d03aa2d9da1223ed2160f8a32c7671e32c9dc1ceef7c5"} Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.004450 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.009339 4904 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.048428 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-874bc5bd9-jqhms"] Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.080931 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.113641 4904 scope.go:117] "RemoveContainer" containerID="dab00463f4f9f87f8e5fd9430d330e67d0b76a0c859df4d3abe75d67fbcfaf80" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.119230 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.202442 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.240241 4904 scope.go:117] "RemoveContainer" containerID="ab57ab2fb1dceb375552d44e51e81505405e2d336ed8cb4e17c1612f7bfadba4" Feb 14 11:26:38 crc kubenswrapper[4904]: E0214 11:26:38.246330 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab57ab2fb1dceb375552d44e51e81505405e2d336ed8cb4e17c1612f7bfadba4\": container with ID starting with ab57ab2fb1dceb375552d44e51e81505405e2d336ed8cb4e17c1612f7bfadba4 not found: ID does not exist" containerID="ab57ab2fb1dceb375552d44e51e81505405e2d336ed8cb4e17c1612f7bfadba4" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.246388 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab57ab2fb1dceb375552d44e51e81505405e2d336ed8cb4e17c1612f7bfadba4"} err="failed to get container status \"ab57ab2fb1dceb375552d44e51e81505405e2d336ed8cb4e17c1612f7bfadba4\": rpc error: code = NotFound desc = could not find container \"ab57ab2fb1dceb375552d44e51e81505405e2d336ed8cb4e17c1612f7bfadba4\": container with ID starting with ab57ab2fb1dceb375552d44e51e81505405e2d336ed8cb4e17c1612f7bfadba4 not found: ID does not exist" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.246420 4904 scope.go:117] "RemoveContainer" containerID="dab00463f4f9f87f8e5fd9430d330e67d0b76a0c859df4d3abe75d67fbcfaf80" Feb 14 11:26:38 crc kubenswrapper[4904]: E0214 11:26:38.250735 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dab00463f4f9f87f8e5fd9430d330e67d0b76a0c859df4d3abe75d67fbcfaf80\": container with ID starting with dab00463f4f9f87f8e5fd9430d330e67d0b76a0c859df4d3abe75d67fbcfaf80 not found: ID does not exist" containerID="dab00463f4f9f87f8e5fd9430d330e67d0b76a0c859df4d3abe75d67fbcfaf80" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.250791 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dab00463f4f9f87f8e5fd9430d330e67d0b76a0c859df4d3abe75d67fbcfaf80"} err="failed to get container status \"dab00463f4f9f87f8e5fd9430d330e67d0b76a0c859df4d3abe75d67fbcfaf80\": rpc error: code = NotFound desc = could not find container \"dab00463f4f9f87f8e5fd9430d330e67d0b76a0c859df4d3abe75d67fbcfaf80\": container with ID starting with dab00463f4f9f87f8e5fd9430d330e67d0b76a0c859df4d3abe75d67fbcfaf80 not found: ID does not exist" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.250818 4904 scope.go:117] "RemoveContainer" containerID="ab57ab2fb1dceb375552d44e51e81505405e2d336ed8cb4e17c1612f7bfadba4" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.250941 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:26:38 crc kubenswrapper[4904]: E0214 11:26:38.251346 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd725c16-a4a1-4dca-b7dc-a159d272f4b0" containerName="registry-server" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.251368 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd725c16-a4a1-4dca-b7dc-a159d272f4b0" containerName="registry-server" Feb 14 11:26:38 crc kubenswrapper[4904]: E0214 11:26:38.251389 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd725c16-a4a1-4dca-b7dc-a159d272f4b0" containerName="extract-content" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.251396 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd725c16-a4a1-4dca-b7dc-a159d272f4b0" containerName="extract-content" Feb 14 11:26:38 crc kubenswrapper[4904]: E0214 11:26:38.251413 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52998ca2-bb09-4f2f-93ce-8e759c30e420" containerName="glance-httpd" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.251421 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="52998ca2-bb09-4f2f-93ce-8e759c30e420" containerName="glance-httpd" Feb 14 11:26:38 crc kubenswrapper[4904]: E0214 11:26:38.251437 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd725c16-a4a1-4dca-b7dc-a159d272f4b0" containerName="extract-utilities" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.251444 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd725c16-a4a1-4dca-b7dc-a159d272f4b0" containerName="extract-utilities" Feb 14 11:26:38 crc kubenswrapper[4904]: E0214 11:26:38.251453 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7" containerName="glance-httpd" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.251460 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7" containerName="glance-httpd" Feb 14 11:26:38 crc kubenswrapper[4904]: E0214 11:26:38.251480 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7" containerName="glance-log" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.251488 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7" containerName="glance-log" Feb 14 11:26:38 crc kubenswrapper[4904]: E0214 11:26:38.251508 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52998ca2-bb09-4f2f-93ce-8e759c30e420" containerName="glance-log" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.251518 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="52998ca2-bb09-4f2f-93ce-8e759c30e420" containerName="glance-log" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.257900 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab57ab2fb1dceb375552d44e51e81505405e2d336ed8cb4e17c1612f7bfadba4"} err="failed to get container status \"ab57ab2fb1dceb375552d44e51e81505405e2d336ed8cb4e17c1612f7bfadba4\": rpc error: code = NotFound desc = could not find container \"ab57ab2fb1dceb375552d44e51e81505405e2d336ed8cb4e17c1612f7bfadba4\": container with ID starting with ab57ab2fb1dceb375552d44e51e81505405e2d336ed8cb4e17c1612f7bfadba4 not found: ID does not exist" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.257949 4904 scope.go:117] "RemoveContainer" containerID="dab00463f4f9f87f8e5fd9430d330e67d0b76a0c859df4d3abe75d67fbcfaf80" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.265344 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dab00463f4f9f87f8e5fd9430d330e67d0b76a0c859df4d3abe75d67fbcfaf80"} err="failed to get container status \"dab00463f4f9f87f8e5fd9430d330e67d0b76a0c859df4d3abe75d67fbcfaf80\": rpc error: code = NotFound desc = could not find container \"dab00463f4f9f87f8e5fd9430d330e67d0b76a0c859df4d3abe75d67fbcfaf80\": container with ID starting with dab00463f4f9f87f8e5fd9430d330e67d0b76a0c859df4d3abe75d67fbcfaf80 not found: ID does not exist" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.265396 4904 scope.go:117] "RemoveContainer" containerID="6b8569215b127848980d65bb56dbb8592e5097497d0e15a1b3593b952c0764d3" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.272052 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7" containerName="glance-httpd" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.272108 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="52998ca2-bb09-4f2f-93ce-8e759c30e420" containerName="glance-httpd" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.272127 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7" containerName="glance-log" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.272139 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd725c16-a4a1-4dca-b7dc-a159d272f4b0" containerName="registry-server" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.272160 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="52998ca2-bb09-4f2f-93ce-8e759c30e420" containerName="glance-log" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.273486 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.279562 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.280026 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.280288 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qlfk8" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.280474 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.293376 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.318712 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-687657b496-tz596"] Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.321592 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.325132 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.345659 4904 scope.go:117] "RemoveContainer" containerID="ab52531d979b58f2b6af5732827f39d98c54f479d4f0eb2695e50e4284f0359e" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.345781 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.385581 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.387513 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.392641 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.393599 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.406577 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.421458 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-687657b496-tz596"] Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.427579 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.427727 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-logs\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.427752 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvccp\" (UniqueName: \"kubernetes.io/projected/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-kube-api-access-fvccp\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.427793 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.427813 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.427850 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.427941 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.427992 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.428345 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:26:38 crc kubenswrapper[4904]: E0214 11:26:38.428827 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data glance httpd-run kube-api-access-6clq8 logs public-tls-certs scripts], unattached volumes=[], failed to process volumes=[combined-ca-bundle config-data glance httpd-run kube-api-access-6clq8 logs public-tls-certs scripts]: context canceled" pod="openstack/glance-default-external-api-0" podUID="8222874b-39c2-49ff-9592-26e35d8e8a0d" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.435946 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-569f4ff6dc-kvsmj"] Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.446500 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6bc9d549b8-hdn6n"] Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.461240 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6bc9d549b8-hdn6n"] Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.461343 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.490176 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:26:38 crc kubenswrapper[4904]: E0214 11:26:38.490853 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-fvccp logs scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-internal-api-0" podUID="409cfcd0-8340-45cf-a309-fd9f5f8e65d6" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.530920 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8222874b-39c2-49ff-9592-26e35d8e8a0d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.530964 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-horizon-tls-certs\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531001 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531020 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-logs\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531037 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8222874b-39c2-49ff-9592-26e35d8e8a0d-logs\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531062 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531082 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531101 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x79gg\" (UniqueName: \"kubernetes.io/projected/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-kube-api-access-x79gg\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531117 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-horizon-secret-key\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531147 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-scripts\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531163 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-scripts\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531183 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531212 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6clq8\" (UniqueName: \"kubernetes.io/projected/8222874b-39c2-49ff-9592-26e35d8e8a0d-kube-api-access-6clq8\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531231 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-logs\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531247 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvccp\" (UniqueName: \"kubernetes.io/projected/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-kube-api-access-fvccp\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531265 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-combined-ca-bundle\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531294 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531321 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531339 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-config-data\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531355 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531376 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531579 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-config-data\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.531599 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.532443 4904 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.534654 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-logs\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.534962 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.537757 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.544422 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.550322 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.550931 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.568995 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvccp\" (UniqueName: \"kubernetes.io/projected/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-kube-api-access-fvccp\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.581678 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.633692 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8222874b-39c2-49ff-9592-26e35d8e8a0d-logs\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.633758 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.633789 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x79gg\" (UniqueName: \"kubernetes.io/projected/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-kube-api-access-x79gg\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.634138 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-horizon-secret-key\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.634191 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-scripts\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.634215 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-scripts\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.634239 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-config-data\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.634271 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.634320 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-horizon-tls-certs\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.634367 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6clq8\" (UniqueName: \"kubernetes.io/projected/8222874b-39c2-49ff-9592-26e35d8e8a0d-kube-api-access-6clq8\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.634399 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-combined-ca-bundle\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.634521 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.635685 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-scripts\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.640284 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8222874b-39c2-49ff-9592-26e35d8e8a0d-logs\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.641104 4904 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.644942 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-horizon-secret-key\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.645020 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-config-data\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.645054 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5ljr\" (UniqueName: \"kubernetes.io/projected/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-kube-api-access-f5ljr\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.645138 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-config-data\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.645201 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-combined-ca-bundle\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.645260 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8222874b-39c2-49ff-9592-26e35d8e8a0d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.645279 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-horizon-tls-certs\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.645321 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-logs\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.645339 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-scripts\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.645374 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-logs\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.645816 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-logs\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.646601 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.646999 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-config-data\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.651892 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.653078 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-combined-ca-bundle\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.653094 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-scripts\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.655388 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-horizon-secret-key\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.657536 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8222874b-39c2-49ff-9592-26e35d8e8a0d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.660951 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-horizon-tls-certs\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.662539 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-config-data\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.664211 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6clq8\" (UniqueName: \"kubernetes.io/projected/8222874b-39c2-49ff-9592-26e35d8e8a0d-kube-api-access-6clq8\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.677003 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x79gg\" (UniqueName: \"kubernetes.io/projected/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-kube-api-access-x79gg\") pod \"horizon-687657b496-tz596\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.685191 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.746625 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-horizon-secret-key\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.746948 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5ljr\" (UniqueName: \"kubernetes.io/projected/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-kube-api-access-f5ljr\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.747016 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-combined-ca-bundle\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.747047 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-logs\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.747068 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-scripts\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.747115 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-config-data\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.747136 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-horizon-tls-certs\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.751879 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-logs\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.760863 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-horizon-tls-certs\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.765802 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-horizon-secret-key\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.771172 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-scripts\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.773548 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-combined-ca-bundle\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.774074 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-config-data\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.774399 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5ljr\" (UniqueName: \"kubernetes.io/projected/cfb252c0-ffb2-4d87-9c4e-b56cc2070d71-kube-api-access-f5ljr\") pod \"horizon-6bc9d549b8-hdn6n\" (UID: \"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71\") " pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.779519 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:26:38 crc kubenswrapper[4904]: I0214 11:26:38.948503 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-687657b496-tz596" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.026876 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.026826 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.047076 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.107456 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.163508 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-internal-tls-certs\") pod \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.163592 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-logs\") pod \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.163641 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.163658 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-httpd-run\") pod \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.163690 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvccp\" (UniqueName: \"kubernetes.io/projected/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-kube-api-access-fvccp\") pod \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.163729 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-scripts\") pod \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.163796 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-combined-ca-bundle\") pod \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.163908 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-config-data\") pod \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\" (UID: \"409cfcd0-8340-45cf-a309-fd9f5f8e65d6\") " Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.166063 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "409cfcd0-8340-45cf-a309-fd9f5f8e65d6" (UID: "409cfcd0-8340-45cf-a309-fd9f5f8e65d6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.171631 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-config-data" (OuterVolumeSpecName: "config-data") pod "409cfcd0-8340-45cf-a309-fd9f5f8e65d6" (UID: "409cfcd0-8340-45cf-a309-fd9f5f8e65d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.172061 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-logs" (OuterVolumeSpecName: "logs") pod "409cfcd0-8340-45cf-a309-fd9f5f8e65d6" (UID: "409cfcd0-8340-45cf-a309-fd9f5f8e65d6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.178121 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "409cfcd0-8340-45cf-a309-fd9f5f8e65d6" (UID: "409cfcd0-8340-45cf-a309-fd9f5f8e65d6"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.179253 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "409cfcd0-8340-45cf-a309-fd9f5f8e65d6" (UID: "409cfcd0-8340-45cf-a309-fd9f5f8e65d6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.180482 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-kube-api-access-fvccp" (OuterVolumeSpecName: "kube-api-access-fvccp") pod "409cfcd0-8340-45cf-a309-fd9f5f8e65d6" (UID: "409cfcd0-8340-45cf-a309-fd9f5f8e65d6"). InnerVolumeSpecName "kube-api-access-fvccp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.180399 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "409cfcd0-8340-45cf-a309-fd9f5f8e65d6" (UID: "409cfcd0-8340-45cf-a309-fd9f5f8e65d6"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.180782 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-scripts" (OuterVolumeSpecName: "scripts") pod "409cfcd0-8340-45cf-a309-fd9f5f8e65d6" (UID: "409cfcd0-8340-45cf-a309-fd9f5f8e65d6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.268354 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"8222874b-39c2-49ff-9592-26e35d8e8a0d\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.268430 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8222874b-39c2-49ff-9592-26e35d8e8a0d-httpd-run\") pod \"8222874b-39c2-49ff-9592-26e35d8e8a0d\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.268447 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8222874b-39c2-49ff-9592-26e35d8e8a0d-logs\") pod \"8222874b-39c2-49ff-9592-26e35d8e8a0d\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.268495 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-scripts\") pod \"8222874b-39c2-49ff-9592-26e35d8e8a0d\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.268525 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6clq8\" (UniqueName: \"kubernetes.io/projected/8222874b-39c2-49ff-9592-26e35d8e8a0d-kube-api-access-6clq8\") pod \"8222874b-39c2-49ff-9592-26e35d8e8a0d\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.268587 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-combined-ca-bundle\") pod \"8222874b-39c2-49ff-9592-26e35d8e8a0d\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.268633 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-public-tls-certs\") pod \"8222874b-39c2-49ff-9592-26e35d8e8a0d\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.268659 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-config-data\") pod \"8222874b-39c2-49ff-9592-26e35d8e8a0d\" (UID: \"8222874b-39c2-49ff-9592-26e35d8e8a0d\") " Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.269090 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.269109 4904 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.269119 4904 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-logs\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.269137 4904 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.269147 4904 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.269155 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvccp\" (UniqueName: \"kubernetes.io/projected/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-kube-api-access-fvccp\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.269163 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.269170 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/409cfcd0-8340-45cf-a309-fd9f5f8e65d6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.270128 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8222874b-39c2-49ff-9592-26e35d8e8a0d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8222874b-39c2-49ff-9592-26e35d8e8a0d" (UID: "8222874b-39c2-49ff-9592-26e35d8e8a0d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.270210 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8222874b-39c2-49ff-9592-26e35d8e8a0d-logs" (OuterVolumeSpecName: "logs") pod "8222874b-39c2-49ff-9592-26e35d8e8a0d" (UID: "8222874b-39c2-49ff-9592-26e35d8e8a0d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.275455 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-scripts" (OuterVolumeSpecName: "scripts") pod "8222874b-39c2-49ff-9592-26e35d8e8a0d" (UID: "8222874b-39c2-49ff-9592-26e35d8e8a0d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.276412 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "8222874b-39c2-49ff-9592-26e35d8e8a0d" (UID: "8222874b-39c2-49ff-9592-26e35d8e8a0d"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.276505 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8222874b-39c2-49ff-9592-26e35d8e8a0d-kube-api-access-6clq8" (OuterVolumeSpecName: "kube-api-access-6clq8") pod "8222874b-39c2-49ff-9592-26e35d8e8a0d" (UID: "8222874b-39c2-49ff-9592-26e35d8e8a0d"). InnerVolumeSpecName "kube-api-access-6clq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.276542 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8222874b-39c2-49ff-9592-26e35d8e8a0d" (UID: "8222874b-39c2-49ff-9592-26e35d8e8a0d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.304463 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8222874b-39c2-49ff-9592-26e35d8e8a0d" (UID: "8222874b-39c2-49ff-9592-26e35d8e8a0d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.304515 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-config-data" (OuterVolumeSpecName: "config-data") pod "8222874b-39c2-49ff-9592-26e35d8e8a0d" (UID: "8222874b-39c2-49ff-9592-26e35d8e8a0d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.339471 4904 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.371064 4904 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.371116 4904 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.371149 4904 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8222874b-39c2-49ff-9592-26e35d8e8a0d-logs\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.371159 4904 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8222874b-39c2-49ff-9592-26e35d8e8a0d-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.371167 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.371175 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6clq8\" (UniqueName: \"kubernetes.io/projected/8222874b-39c2-49ff-9592-26e35d8e8a0d-kube-api-access-6clq8\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.371184 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.371191 4904 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.371198 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8222874b-39c2-49ff-9592-26e35d8e8a0d-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.392683 4904 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.448218 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-lrm4d" podUID="b2bcaf28-8b82-460c-9535-b9404a3d19a9" containerName="registry-server" probeResult="failure" output=< Feb 14 11:26:39 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 11:26:39 crc kubenswrapper[4904]: > Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.481905 4904 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.501370 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6bc9d549b8-hdn6n"] Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.629613 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-687657b496-tz596"] Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.814066 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.882438 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52998ca2-bb09-4f2f-93ce-8e759c30e420" path="/var/lib/kubelet/pods/52998ca2-bb09-4f2f-93ce-8e759c30e420/volumes" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.883257 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7" path="/var/lib/kubelet/pods/f002abcb-4a1a-4cd4-8ecc-7e42f960a5e7/volumes" Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.891413 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-dp9lt"] Feb 14 11:26:39 crc kubenswrapper[4904]: I0214 11:26:39.891659 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-dp9lt" podUID="4d6d0a8f-2600-47c4-95a0-10f308e493b5" containerName="dnsmasq-dns" containerID="cri-o://68bcf29edaf47bf7889aa5762ca5390fc72fa16206dd5f0776a150a1958a2278" gracePeriod=10 Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.083204 4904 generic.go:334] "Generic (PLEG): container finished" podID="51d73444-44fa-47bf-b8a4-d5c46b0cf80a" containerID="6bb8ce5e37d9faa7fb04dbc186646c0baff6a994d8e2887ed71781d05f2a5b2d" exitCode=0 Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.083303 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.083938 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7bjcl" event={"ID":"51d73444-44fa-47bf-b8a4-d5c46b0cf80a","Type":"ContainerDied","Data":"6bb8ce5e37d9faa7fb04dbc186646c0baff6a994d8e2887ed71781d05f2a5b2d"} Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.084054 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.243058 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.262478 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.290487 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.292921 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.304345 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qlfk8" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.304512 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.305163 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.305406 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.338890 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.354736 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.405429 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.409765 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-scripts\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.409846 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.409905 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.409941 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-logs\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.409973 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.410059 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-config-data\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.410432 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qll2q\" (UniqueName: \"kubernetes.io/projected/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-kube-api-access-qll2q\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.410459 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.414554 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.416099 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.421374 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.421543 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.498541 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.511562 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-scripts\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.511615 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.511641 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.511676 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.511714 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.511738 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-logs\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.511756 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.511778 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.511807 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.511993 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-config-data\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.512039 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qll2q\" (UniqueName: \"kubernetes.io/projected/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-kube-api-access-qll2q\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.512056 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.512075 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2a959004-612f-45ee-82ee-9d1fe7af7eba-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.512125 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a959004-612f-45ee-82ee-9d1fe7af7eba-logs\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.512154 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.512196 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsjkv\" (UniqueName: \"kubernetes.io/projected/2a959004-612f-45ee-82ee-9d1fe7af7eba-kube-api-access-dsjkv\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.512816 4904 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.512930 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.513214 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-logs\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.521478 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.532792 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-scripts\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.544956 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.547424 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-config-data\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.549059 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qll2q\" (UniqueName: \"kubernetes.io/projected/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-kube-api-access-qll2q\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.598214 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.615137 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.615184 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsjkv\" (UniqueName: \"kubernetes.io/projected/2a959004-612f-45ee-82ee-9d1fe7af7eba-kube-api-access-dsjkv\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.615238 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.615270 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.615331 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.615353 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.615397 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2a959004-612f-45ee-82ee-9d1fe7af7eba-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.615426 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a959004-612f-45ee-82ee-9d1fe7af7eba-logs\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.615803 4904 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.615916 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a959004-612f-45ee-82ee-9d1fe7af7eba-logs\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.616637 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2a959004-612f-45ee-82ee-9d1fe7af7eba-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.622508 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.623007 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.623869 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.625324 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.632516 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsjkv\" (UniqueName: \"kubernetes.io/projected/2a959004-612f-45ee-82ee-9d1fe7af7eba-kube-api-access-dsjkv\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.639819 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.681867 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:26:40 crc kubenswrapper[4904]: I0214 11:26:40.746052 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 14 11:26:41 crc kubenswrapper[4904]: I0214 11:26:41.102906 4904 generic.go:334] "Generic (PLEG): container finished" podID="4d6d0a8f-2600-47c4-95a0-10f308e493b5" containerID="68bcf29edaf47bf7889aa5762ca5390fc72fa16206dd5f0776a150a1958a2278" exitCode=0 Feb 14 11:26:41 crc kubenswrapper[4904]: I0214 11:26:41.102985 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-dp9lt" event={"ID":"4d6d0a8f-2600-47c4-95a0-10f308e493b5","Type":"ContainerDied","Data":"68bcf29edaf47bf7889aa5762ca5390fc72fa16206dd5f0776a150a1958a2278"} Feb 14 11:26:41 crc kubenswrapper[4904]: I0214 11:26:41.850778 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="409cfcd0-8340-45cf-a309-fd9f5f8e65d6" path="/var/lib/kubelet/pods/409cfcd0-8340-45cf-a309-fd9f5f8e65d6/volumes" Feb 14 11:26:41 crc kubenswrapper[4904]: I0214 11:26:41.851623 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8222874b-39c2-49ff-9592-26e35d8e8a0d" path="/var/lib/kubelet/pods/8222874b-39c2-49ff-9592-26e35d8e8a0d/volumes" Feb 14 11:26:42 crc kubenswrapper[4904]: I0214 11:26:42.753437 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-dp9lt" podUID="4d6d0a8f-2600-47c4-95a0-10f308e493b5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.109:5353: connect: connection refused" Feb 14 11:26:43 crc kubenswrapper[4904]: I0214 11:26:43.129376 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-687657b496-tz596" event={"ID":"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393","Type":"ContainerStarted","Data":"b8527c01a195ffb25243d7944cf0993c62d9b22869ca946577bf35f1a1a95a82"} Feb 14 11:26:43 crc kubenswrapper[4904]: I0214 11:26:43.203474 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8v2xr" podUID="a9fd9edc-0058-417c-baa6-9fbff5035af5" containerName="registry-server" probeResult="failure" output=< Feb 14 11:26:43 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 11:26:43 crc kubenswrapper[4904]: > Feb 14 11:26:47 crc kubenswrapper[4904]: I0214 11:26:47.753472 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-dp9lt" podUID="4d6d0a8f-2600-47c4-95a0-10f308e493b5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.109:5353: connect: connection refused" Feb 14 11:26:48 crc kubenswrapper[4904]: I0214 11:26:48.390497 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lrm4d" Feb 14 11:26:48 crc kubenswrapper[4904]: I0214 11:26:48.452945 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lrm4d" Feb 14 11:26:49 crc kubenswrapper[4904]: I0214 11:26:49.212728 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lrm4d"] Feb 14 11:26:50 crc kubenswrapper[4904]: I0214 11:26:50.187884 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lrm4d" podUID="b2bcaf28-8b82-460c-9535-b9404a3d19a9" containerName="registry-server" containerID="cri-o://db123fad5e9b0e82052c07b64897a7abfc8be7140e0070fcf018b76330f9cb21" gracePeriod=2 Feb 14 11:26:51 crc kubenswrapper[4904]: I0214 11:26:51.198595 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lrm4d" event={"ID":"b2bcaf28-8b82-460c-9535-b9404a3d19a9","Type":"ContainerDied","Data":"db123fad5e9b0e82052c07b64897a7abfc8be7140e0070fcf018b76330f9cb21"} Feb 14 11:26:51 crc kubenswrapper[4904]: I0214 11:26:51.198415 4904 generic.go:334] "Generic (PLEG): container finished" podID="b2bcaf28-8b82-460c-9535-b9404a3d19a9" containerID="db123fad5e9b0e82052c07b64897a7abfc8be7140e0070fcf018b76330f9cb21" exitCode=0 Feb 14 11:26:52 crc kubenswrapper[4904]: E0214 11:26:52.308911 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 14 11:26:52 crc kubenswrapper[4904]: E0214 11:26:52.309317 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n568h669h5c5h5c8h77h5d4h5d9h5b7h68dh598h8dh676h85h54ch88h5d4h54bh5c8h578h5b5hffhf9hbch57hd9h686h649h5f4h5bch669h66bh5c5q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tbx78,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6b79d987b9-h8nmk_openstack(97bc1e53-5edf-43fc-8aca-0dee904c86eb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:26:52 crc kubenswrapper[4904]: E0214 11:26:52.312118 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-6b79d987b9-h8nmk" podUID="97bc1e53-5edf-43fc-8aca-0dee904c86eb" Feb 14 11:26:52 crc kubenswrapper[4904]: E0214 11:26:52.330667 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 14 11:26:52 crc kubenswrapper[4904]: E0214 11:26:52.330945 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5h89h544h674h575h5dch599h68bh5fh7bh668h555h98h8dh85h58dh575hbchddh5c4h686h645h5fh5c5h689h59ch697h5f4h57bh5cbhcbh87q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v5q25,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-569f4ff6dc-kvsmj_openstack(fa07aa03-b19b-42ba-9834-0ce67c1df067): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:26:52 crc kubenswrapper[4904]: E0214 11:26:52.336493 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-569f4ff6dc-kvsmj" podUID="fa07aa03-b19b-42ba-9834-0ce67c1df067" Feb 14 11:26:52 crc kubenswrapper[4904]: I0214 11:26:52.753867 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-dp9lt" podUID="4d6d0a8f-2600-47c4-95a0-10f308e493b5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.109:5353: connect: connection refused" Feb 14 11:26:52 crc kubenswrapper[4904]: I0214 11:26:52.754008 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:26:53 crc kubenswrapper[4904]: I0214 11:26:53.174296 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8v2xr" podUID="a9fd9edc-0058-417c-baa6-9fbff5035af5" containerName="registry-server" probeResult="failure" output=< Feb 14 11:26:53 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 11:26:53 crc kubenswrapper[4904]: > Feb 14 11:26:53 crc kubenswrapper[4904]: I0214 11:26:53.216220 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bc9d549b8-hdn6n" event={"ID":"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71","Type":"ContainerStarted","Data":"a26d15ab68f88add510cfd0f82c764373df59b26a6f7069fa26911eba91b464a"} Feb 14 11:26:55 crc kubenswrapper[4904]: E0214 11:26:55.040996 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Feb 14 11:26:55 crc kubenswrapper[4904]: E0214 11:26:55.041284 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n577hbbh56dh88h54fhd4h8dh8dh694h5ffhfbh5fdh7h5fdh567hd4h87h646h687h64h67h5cfh685h586h674h8h59ch588h55fhf8h5bfh58bq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xgmqf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(16339330-a7b5-471a-ad7e-1a8d4627c3e2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:26:55 crc kubenswrapper[4904]: E0214 11:26:55.056187 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 14 11:26:55 crc kubenswrapper[4904]: E0214 11:26:55.056354 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n699h65bhf6h5bbh66fh56dh548hfch5bh79h56fh54dh695h5d4h665h554hbh5bch5c4h5b9hcbh65ch57fhcchb7hcbh86hddh685h5bh5h97q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bd97l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-874bc5bd9-jqhms_openstack(0c6ef645-08d7-4bd0-aa8a-8181bae77b28): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:26:55 crc kubenswrapper[4904]: E0214 11:26:55.058901 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-874bc5bd9-jqhms" podUID="0c6ef645-08d7-4bd0-aa8a-8181bae77b28" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.141051 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.144726 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-569f4ff6dc-kvsmj" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.156172 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b79d987b9-h8nmk" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.204165 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-fernet-keys\") pod \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.204269 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkvld\" (UniqueName: \"kubernetes.io/projected/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-kube-api-access-qkvld\") pod \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.204316 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-config-data\") pod \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.204396 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-combined-ca-bundle\") pod \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.204497 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-scripts\") pod \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.204525 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-credential-keys\") pod \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\" (UID: \"51d73444-44fa-47bf-b8a4-d5c46b0cf80a\") " Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.215245 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-scripts" (OuterVolumeSpecName: "scripts") pod "51d73444-44fa-47bf-b8a4-d5c46b0cf80a" (UID: "51d73444-44fa-47bf-b8a4-d5c46b0cf80a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.217771 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "51d73444-44fa-47bf-b8a4-d5c46b0cf80a" (UID: "51d73444-44fa-47bf-b8a4-d5c46b0cf80a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.221336 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-kube-api-access-qkvld" (OuterVolumeSpecName: "kube-api-access-qkvld") pod "51d73444-44fa-47bf-b8a4-d5c46b0cf80a" (UID: "51d73444-44fa-47bf-b8a4-d5c46b0cf80a"). InnerVolumeSpecName "kube-api-access-qkvld". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.252192 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51d73444-44fa-47bf-b8a4-d5c46b0cf80a" (UID: "51d73444-44fa-47bf-b8a4-d5c46b0cf80a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.258818 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "51d73444-44fa-47bf-b8a4-d5c46b0cf80a" (UID: "51d73444-44fa-47bf-b8a4-d5c46b0cf80a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.272938 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-569f4ff6dc-kvsmj" event={"ID":"fa07aa03-b19b-42ba-9834-0ce67c1df067","Type":"ContainerDied","Data":"5b326289aa04248f581504c47df3509463c6e29f3bbd7ed3ffc4925d788fb974"} Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.273206 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-569f4ff6dc-kvsmj" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.297152 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b79d987b9-h8nmk" event={"ID":"97bc1e53-5edf-43fc-8aca-0dee904c86eb","Type":"ContainerDied","Data":"8d966d7fb7d3fd39ec5fa4d4980bb6d767da93884c37a8955b66bb3bbc7130c0"} Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.297234 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b79d987b9-h8nmk" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.297419 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-config-data" (OuterVolumeSpecName: "config-data") pod "51d73444-44fa-47bf-b8a4-d5c46b0cf80a" (UID: "51d73444-44fa-47bf-b8a4-d5c46b0cf80a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.308318 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7bjcl" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.308354 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7bjcl" event={"ID":"51d73444-44fa-47bf-b8a4-d5c46b0cf80a","Type":"ContainerDied","Data":"59c338cdaf7c0b4c5dcee4c5efb50931c86534e944cfcedccaad2601ebf22c65"} Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.308383 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59c338cdaf7c0b4c5dcee4c5efb50931c86534e944cfcedccaad2601ebf22c65" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.312033 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97bc1e53-5edf-43fc-8aca-0dee904c86eb-config-data\") pod \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.312119 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa07aa03-b19b-42ba-9834-0ce67c1df067-logs\") pod \"fa07aa03-b19b-42ba-9834-0ce67c1df067\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.312165 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbx78\" (UniqueName: \"kubernetes.io/projected/97bc1e53-5edf-43fc-8aca-0dee904c86eb-kube-api-access-tbx78\") pod \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.312201 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/97bc1e53-5edf-43fc-8aca-0dee904c86eb-horizon-secret-key\") pod \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.312248 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fa07aa03-b19b-42ba-9834-0ce67c1df067-horizon-secret-key\") pod \"fa07aa03-b19b-42ba-9834-0ce67c1df067\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.312334 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa07aa03-b19b-42ba-9834-0ce67c1df067-scripts\") pod \"fa07aa03-b19b-42ba-9834-0ce67c1df067\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.312408 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fa07aa03-b19b-42ba-9834-0ce67c1df067-config-data\") pod \"fa07aa03-b19b-42ba-9834-0ce67c1df067\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.312442 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5q25\" (UniqueName: \"kubernetes.io/projected/fa07aa03-b19b-42ba-9834-0ce67c1df067-kube-api-access-v5q25\") pod \"fa07aa03-b19b-42ba-9834-0ce67c1df067\" (UID: \"fa07aa03-b19b-42ba-9834-0ce67c1df067\") " Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.312512 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97bc1e53-5edf-43fc-8aca-0dee904c86eb-logs\") pod \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.312532 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97bc1e53-5edf-43fc-8aca-0dee904c86eb-scripts\") pod \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\" (UID: \"97bc1e53-5edf-43fc-8aca-0dee904c86eb\") " Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.312896 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.312913 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.313286 4904 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.313299 4904 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.313312 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkvld\" (UniqueName: \"kubernetes.io/projected/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-kube-api-access-qkvld\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.313324 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51d73444-44fa-47bf-b8a4-d5c46b0cf80a-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.313517 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97bc1e53-5edf-43fc-8aca-0dee904c86eb-config-data" (OuterVolumeSpecName: "config-data") pod "97bc1e53-5edf-43fc-8aca-0dee904c86eb" (UID: "97bc1e53-5edf-43fc-8aca-0dee904c86eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.314145 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97bc1e53-5edf-43fc-8aca-0dee904c86eb-scripts" (OuterVolumeSpecName: "scripts") pod "97bc1e53-5edf-43fc-8aca-0dee904c86eb" (UID: "97bc1e53-5edf-43fc-8aca-0dee904c86eb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.314590 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa07aa03-b19b-42ba-9834-0ce67c1df067-scripts" (OuterVolumeSpecName: "scripts") pod "fa07aa03-b19b-42ba-9834-0ce67c1df067" (UID: "fa07aa03-b19b-42ba-9834-0ce67c1df067"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.315940 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa07aa03-b19b-42ba-9834-0ce67c1df067-logs" (OuterVolumeSpecName: "logs") pod "fa07aa03-b19b-42ba-9834-0ce67c1df067" (UID: "fa07aa03-b19b-42ba-9834-0ce67c1df067"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.315976 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa07aa03-b19b-42ba-9834-0ce67c1df067-config-data" (OuterVolumeSpecName: "config-data") pod "fa07aa03-b19b-42ba-9834-0ce67c1df067" (UID: "fa07aa03-b19b-42ba-9834-0ce67c1df067"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.316711 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97bc1e53-5edf-43fc-8aca-0dee904c86eb-logs" (OuterVolumeSpecName: "logs") pod "97bc1e53-5edf-43fc-8aca-0dee904c86eb" (UID: "97bc1e53-5edf-43fc-8aca-0dee904c86eb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.325814 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97bc1e53-5edf-43fc-8aca-0dee904c86eb-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "97bc1e53-5edf-43fc-8aca-0dee904c86eb" (UID: "97bc1e53-5edf-43fc-8aca-0dee904c86eb"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.326222 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa07aa03-b19b-42ba-9834-0ce67c1df067-kube-api-access-v5q25" (OuterVolumeSpecName: "kube-api-access-v5q25") pod "fa07aa03-b19b-42ba-9834-0ce67c1df067" (UID: "fa07aa03-b19b-42ba-9834-0ce67c1df067"). InnerVolumeSpecName "kube-api-access-v5q25". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.327357 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa07aa03-b19b-42ba-9834-0ce67c1df067-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "fa07aa03-b19b-42ba-9834-0ce67c1df067" (UID: "fa07aa03-b19b-42ba-9834-0ce67c1df067"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.327919 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97bc1e53-5edf-43fc-8aca-0dee904c86eb-kube-api-access-tbx78" (OuterVolumeSpecName: "kube-api-access-tbx78") pod "97bc1e53-5edf-43fc-8aca-0dee904c86eb" (UID: "97bc1e53-5edf-43fc-8aca-0dee904c86eb"). InnerVolumeSpecName "kube-api-access-tbx78". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.419010 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa07aa03-b19b-42ba-9834-0ce67c1df067-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.419041 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fa07aa03-b19b-42ba-9834-0ce67c1df067-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.419050 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5q25\" (UniqueName: \"kubernetes.io/projected/fa07aa03-b19b-42ba-9834-0ce67c1df067-kube-api-access-v5q25\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.419062 4904 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97bc1e53-5edf-43fc-8aca-0dee904c86eb-logs\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.419070 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97bc1e53-5edf-43fc-8aca-0dee904c86eb-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.419079 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97bc1e53-5edf-43fc-8aca-0dee904c86eb-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.419088 4904 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa07aa03-b19b-42ba-9834-0ce67c1df067-logs\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.419097 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbx78\" (UniqueName: \"kubernetes.io/projected/97bc1e53-5edf-43fc-8aca-0dee904c86eb-kube-api-access-tbx78\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.419105 4904 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/97bc1e53-5edf-43fc-8aca-0dee904c86eb-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.419112 4904 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fa07aa03-b19b-42ba-9834-0ce67c1df067-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.658615 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-569f4ff6dc-kvsmj"] Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.665693 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-569f4ff6dc-kvsmj"] Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.698894 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6b79d987b9-h8nmk"] Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.713914 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6b79d987b9-h8nmk"] Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.845193 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97bc1e53-5edf-43fc-8aca-0dee904c86eb" path="/var/lib/kubelet/pods/97bc1e53-5edf-43fc-8aca-0dee904c86eb/volumes" Feb 14 11:26:55 crc kubenswrapper[4904]: I0214 11:26:55.845646 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa07aa03-b19b-42ba-9834-0ce67c1df067" path="/var/lib/kubelet/pods/fa07aa03-b19b-42ba-9834-0ce67c1df067/volumes" Feb 14 11:26:56 crc kubenswrapper[4904]: E0214 11:26:56.016747 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Feb 14 11:26:56 crc kubenswrapper[4904]: E0214 11:26:56.016918 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f4djz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-sflms_openstack(6b349fb1-5e71-45ff-88e9-03592bfe2504): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:26:56 crc kubenswrapper[4904]: E0214 11:26:56.021003 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-sflms" podUID="6b349fb1-5e71-45ff-88e9-03592bfe2504" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.242852 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-7bjcl"] Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.252714 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-7bjcl"] Feb 14 11:26:56 crc kubenswrapper[4904]: E0214 11:26:56.322755 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-sflms" podUID="6b349fb1-5e71-45ff-88e9-03592bfe2504" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.340952 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-v6zmt"] Feb 14 11:26:56 crc kubenswrapper[4904]: E0214 11:26:56.341328 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51d73444-44fa-47bf-b8a4-d5c46b0cf80a" containerName="keystone-bootstrap" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.341341 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="51d73444-44fa-47bf-b8a4-d5c46b0cf80a" containerName="keystone-bootstrap" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.341498 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="51d73444-44fa-47bf-b8a4-d5c46b0cf80a" containerName="keystone-bootstrap" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.342813 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.344776 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.345128 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.345293 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.346707 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.346958 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-57x7t" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.373825 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-v6zmt"] Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.445002 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-fernet-keys\") pod \"keystone-bootstrap-v6zmt\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.445045 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2j7s\" (UniqueName: \"kubernetes.io/projected/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-kube-api-access-h2j7s\") pod \"keystone-bootstrap-v6zmt\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.445069 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-config-data\") pod \"keystone-bootstrap-v6zmt\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.445091 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-credential-keys\") pod \"keystone-bootstrap-v6zmt\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.445107 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-combined-ca-bundle\") pod \"keystone-bootstrap-v6zmt\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.445785 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-scripts\") pod \"keystone-bootstrap-v6zmt\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.547823 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-scripts\") pod \"keystone-bootstrap-v6zmt\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.547895 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-fernet-keys\") pod \"keystone-bootstrap-v6zmt\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.547917 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2j7s\" (UniqueName: \"kubernetes.io/projected/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-kube-api-access-h2j7s\") pod \"keystone-bootstrap-v6zmt\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.547934 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-config-data\") pod \"keystone-bootstrap-v6zmt\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.547957 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-credential-keys\") pod \"keystone-bootstrap-v6zmt\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.548712 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-combined-ca-bundle\") pod \"keystone-bootstrap-v6zmt\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.553541 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-fernet-keys\") pod \"keystone-bootstrap-v6zmt\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.554480 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-config-data\") pod \"keystone-bootstrap-v6zmt\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.557097 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-scripts\") pod \"keystone-bootstrap-v6zmt\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.558999 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-credential-keys\") pod \"keystone-bootstrap-v6zmt\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.562690 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2j7s\" (UniqueName: \"kubernetes.io/projected/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-kube-api-access-h2j7s\") pod \"keystone-bootstrap-v6zmt\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.569510 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-combined-ca-bundle\") pod \"keystone-bootstrap-v6zmt\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:56 crc kubenswrapper[4904]: I0214 11:26:56.660042 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:26:57 crc kubenswrapper[4904]: I0214 11:26:57.856333 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51d73444-44fa-47bf-b8a4-d5c46b0cf80a" path="/var/lib/kubelet/pods/51d73444-44fa-47bf-b8a4-d5c46b0cf80a/volumes" Feb 14 11:26:58 crc kubenswrapper[4904]: E0214 11:26:58.329614 4904 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of db123fad5e9b0e82052c07b64897a7abfc8be7140e0070fcf018b76330f9cb21 is running failed: container process not found" containerID="db123fad5e9b0e82052c07b64897a7abfc8be7140e0070fcf018b76330f9cb21" cmd=["grpc_health_probe","-addr=:50051"] Feb 14 11:26:58 crc kubenswrapper[4904]: E0214 11:26:58.330503 4904 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of db123fad5e9b0e82052c07b64897a7abfc8be7140e0070fcf018b76330f9cb21 is running failed: container process not found" containerID="db123fad5e9b0e82052c07b64897a7abfc8be7140e0070fcf018b76330f9cb21" cmd=["grpc_health_probe","-addr=:50051"] Feb 14 11:26:58 crc kubenswrapper[4904]: E0214 11:26:58.332238 4904 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of db123fad5e9b0e82052c07b64897a7abfc8be7140e0070fcf018b76330f9cb21 is running failed: container process not found" containerID="db123fad5e9b0e82052c07b64897a7abfc8be7140e0070fcf018b76330f9cb21" cmd=["grpc_health_probe","-addr=:50051"] Feb 14 11:26:58 crc kubenswrapper[4904]: E0214 11:26:58.332276 4904 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of db123fad5e9b0e82052c07b64897a7abfc8be7140e0070fcf018b76330f9cb21 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-lrm4d" podUID="b2bcaf28-8b82-460c-9535-b9404a3d19a9" containerName="registry-server" Feb 14 11:26:58 crc kubenswrapper[4904]: I0214 11:26:58.352380 4904 generic.go:334] "Generic (PLEG): container finished" podID="2abcf43f-6f3b-4804-83a5-ba42fc66f2ec" containerID="41db167ff7dc3e24f9b2ac1e32c2405e16f5f00ee2611e44e5400ce8318cbd67" exitCode=0 Feb 14 11:26:58 crc kubenswrapper[4904]: I0214 11:26:58.352431 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-t6bn6" event={"ID":"2abcf43f-6f3b-4804-83a5-ba42fc66f2ec","Type":"ContainerDied","Data":"41db167ff7dc3e24f9b2ac1e32c2405e16f5f00ee2611e44e5400ce8318cbd67"} Feb 14 11:27:02 crc kubenswrapper[4904]: I0214 11:27:02.174722 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8v2xr" Feb 14 11:27:02 crc kubenswrapper[4904]: I0214 11:27:02.221136 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8v2xr" Feb 14 11:27:02 crc kubenswrapper[4904]: I0214 11:27:02.412310 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8v2xr"] Feb 14 11:27:02 crc kubenswrapper[4904]: I0214 11:27:02.754424 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-dp9lt" podUID="4d6d0a8f-2600-47c4-95a0-10f308e493b5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.109:5353: i/o timeout" Feb 14 11:27:03 crc kubenswrapper[4904]: I0214 11:27:03.392272 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8v2xr" podUID="a9fd9edc-0058-417c-baa6-9fbff5035af5" containerName="registry-server" containerID="cri-o://522c8e0477ec7ed39163c6dc7ca5c2542ed209a4caa41b362756bc571a8ad301" gracePeriod=2 Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.405955 4904 generic.go:334] "Generic (PLEG): container finished" podID="a9fd9edc-0058-417c-baa6-9fbff5035af5" containerID="522c8e0477ec7ed39163c6dc7ca5c2542ed209a4caa41b362756bc571a8ad301" exitCode=0 Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.406333 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8v2xr" event={"ID":"a9fd9edc-0058-417c-baa6-9fbff5035af5","Type":"ContainerDied","Data":"522c8e0477ec7ed39163c6dc7ca5c2542ed209a4caa41b362756bc571a8ad301"} Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.720731 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.754610 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lrm4d" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.764204 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-874bc5bd9-jqhms" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.770423 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-t6bn6" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.818966 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znh26\" (UniqueName: \"kubernetes.io/projected/4d6d0a8f-2600-47c4-95a0-10f308e493b5-kube-api-access-znh26\") pod \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.819012 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-config\") pod \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.819112 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-ovsdbserver-sb\") pod \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.819141 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-ovsdbserver-nb\") pod \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.819233 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-dns-svc\") pod \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\" (UID: \"4d6d0a8f-2600-47c4-95a0-10f308e493b5\") " Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.881260 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d6d0a8f-2600-47c4-95a0-10f308e493b5-kube-api-access-znh26" (OuterVolumeSpecName: "kube-api-access-znh26") pod "4d6d0a8f-2600-47c4-95a0-10f308e493b5" (UID: "4d6d0a8f-2600-47c4-95a0-10f308e493b5"). InnerVolumeSpecName "kube-api-access-znh26". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.911582 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4d6d0a8f-2600-47c4-95a0-10f308e493b5" (UID: "4d6d0a8f-2600-47c4-95a0-10f308e493b5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.921034 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzt7z\" (UniqueName: \"kubernetes.io/projected/2abcf43f-6f3b-4804-83a5-ba42fc66f2ec-kube-api-access-wzt7z\") pod \"2abcf43f-6f3b-4804-83a5-ba42fc66f2ec\" (UID: \"2abcf43f-6f3b-4804-83a5-ba42fc66f2ec\") " Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.921083 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-logs\") pod \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.921126 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2bcaf28-8b82-460c-9535-b9404a3d19a9-catalog-content\") pod \"b2bcaf28-8b82-460c-9535-b9404a3d19a9\" (UID: \"b2bcaf28-8b82-460c-9535-b9404a3d19a9\") " Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.921269 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bd97l\" (UniqueName: \"kubernetes.io/projected/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-kube-api-access-bd97l\") pod \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.921332 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-horizon-secret-key\") pod \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.921446 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2abcf43f-6f3b-4804-83a5-ba42fc66f2ec-combined-ca-bundle\") pod \"2abcf43f-6f3b-4804-83a5-ba42fc66f2ec\" (UID: \"2abcf43f-6f3b-4804-83a5-ba42fc66f2ec\") " Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.921514 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2abcf43f-6f3b-4804-83a5-ba42fc66f2ec-config\") pod \"2abcf43f-6f3b-4804-83a5-ba42fc66f2ec\" (UID: \"2abcf43f-6f3b-4804-83a5-ba42fc66f2ec\") " Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.921553 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlkbm\" (UniqueName: \"kubernetes.io/projected/b2bcaf28-8b82-460c-9535-b9404a3d19a9-kube-api-access-jlkbm\") pod \"b2bcaf28-8b82-460c-9535-b9404a3d19a9\" (UID: \"b2bcaf28-8b82-460c-9535-b9404a3d19a9\") " Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.921573 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-config-data\") pod \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.921596 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-scripts\") pod \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\" (UID: \"0c6ef645-08d7-4bd0-aa8a-8181bae77b28\") " Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.921640 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2bcaf28-8b82-460c-9535-b9404a3d19a9-utilities\") pod \"b2bcaf28-8b82-460c-9535-b9404a3d19a9\" (UID: \"b2bcaf28-8b82-460c-9535-b9404a3d19a9\") " Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.922028 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-logs" (OuterVolumeSpecName: "logs") pod "0c6ef645-08d7-4bd0-aa8a-8181bae77b28" (UID: "0c6ef645-08d7-4bd0-aa8a-8181bae77b28"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.922869 4904 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.922907 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znh26\" (UniqueName: \"kubernetes.io/projected/4d6d0a8f-2600-47c4-95a0-10f308e493b5-kube-api-access-znh26\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.922917 4904 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-logs\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.923305 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2bcaf28-8b82-460c-9535-b9404a3d19a9-utilities" (OuterVolumeSpecName: "utilities") pod "b2bcaf28-8b82-460c-9535-b9404a3d19a9" (UID: "b2bcaf28-8b82-460c-9535-b9404a3d19a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.924382 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-scripts" (OuterVolumeSpecName: "scripts") pod "0c6ef645-08d7-4bd0-aa8a-8181bae77b28" (UID: "0c6ef645-08d7-4bd0-aa8a-8181bae77b28"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.925239 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-config-data" (OuterVolumeSpecName: "config-data") pod "0c6ef645-08d7-4bd0-aa8a-8181bae77b28" (UID: "0c6ef645-08d7-4bd0-aa8a-8181bae77b28"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.928412 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2bcaf28-8b82-460c-9535-b9404a3d19a9-kube-api-access-jlkbm" (OuterVolumeSpecName: "kube-api-access-jlkbm") pod "b2bcaf28-8b82-460c-9535-b9404a3d19a9" (UID: "b2bcaf28-8b82-460c-9535-b9404a3d19a9"). InnerVolumeSpecName "kube-api-access-jlkbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.929230 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-config" (OuterVolumeSpecName: "config") pod "4d6d0a8f-2600-47c4-95a0-10f308e493b5" (UID: "4d6d0a8f-2600-47c4-95a0-10f308e493b5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.933112 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "0c6ef645-08d7-4bd0-aa8a-8181bae77b28" (UID: "0c6ef645-08d7-4bd0-aa8a-8181bae77b28"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.969086 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2abcf43f-6f3b-4804-83a5-ba42fc66f2ec-kube-api-access-wzt7z" (OuterVolumeSpecName: "kube-api-access-wzt7z") pod "2abcf43f-6f3b-4804-83a5-ba42fc66f2ec" (UID: "2abcf43f-6f3b-4804-83a5-ba42fc66f2ec"). InnerVolumeSpecName "kube-api-access-wzt7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.970014 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4d6d0a8f-2600-47c4-95a0-10f308e493b5" (UID: "4d6d0a8f-2600-47c4-95a0-10f308e493b5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.971244 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-kube-api-access-bd97l" (OuterVolumeSpecName: "kube-api-access-bd97l") pod "0c6ef645-08d7-4bd0-aa8a-8181bae77b28" (UID: "0c6ef645-08d7-4bd0-aa8a-8181bae77b28"). InnerVolumeSpecName "kube-api-access-bd97l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.977153 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4d6d0a8f-2600-47c4-95a0-10f308e493b5" (UID: "4d6d0a8f-2600-47c4-95a0-10f308e493b5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.988355 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2abcf43f-6f3b-4804-83a5-ba42fc66f2ec-config" (OuterVolumeSpecName: "config") pod "2abcf43f-6f3b-4804-83a5-ba42fc66f2ec" (UID: "2abcf43f-6f3b-4804-83a5-ba42fc66f2ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.989051 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2abcf43f-6f3b-4804-83a5-ba42fc66f2ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2abcf43f-6f3b-4804-83a5-ba42fc66f2ec" (UID: "2abcf43f-6f3b-4804-83a5-ba42fc66f2ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:04 crc kubenswrapper[4904]: I0214 11:27:04.990521 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2bcaf28-8b82-460c-9535-b9404a3d19a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2bcaf28-8b82-460c-9535-b9404a3d19a9" (UID: "b2bcaf28-8b82-460c-9535-b9404a3d19a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.024208 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzt7z\" (UniqueName: \"kubernetes.io/projected/2abcf43f-6f3b-4804-83a5-ba42fc66f2ec-kube-api-access-wzt7z\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.024245 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2bcaf28-8b82-460c-9535-b9404a3d19a9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.024254 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.024265 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.024275 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bd97l\" (UniqueName: \"kubernetes.io/projected/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-kube-api-access-bd97l\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.024285 4904 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.024293 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2abcf43f-6f3b-4804-83a5-ba42fc66f2ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.024302 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2abcf43f-6f3b-4804-83a5-ba42fc66f2ec-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.024310 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlkbm\" (UniqueName: \"kubernetes.io/projected/b2bcaf28-8b82-460c-9535-b9404a3d19a9-kube-api-access-jlkbm\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.024318 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.024326 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c6ef645-08d7-4bd0-aa8a-8181bae77b28-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.024333 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d6d0a8f-2600-47c4-95a0-10f308e493b5-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.024341 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2bcaf28-8b82-460c-9535-b9404a3d19a9-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.417750 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lrm4d" event={"ID":"b2bcaf28-8b82-460c-9535-b9404a3d19a9","Type":"ContainerDied","Data":"8f6cf37d10ae646369ca48cc304f7e52812d87fab799e885faa6c4e1a750adfd"} Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.417784 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lrm4d" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.417799 4904 scope.go:117] "RemoveContainer" containerID="db123fad5e9b0e82052c07b64897a7abfc8be7140e0070fcf018b76330f9cb21" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.424580 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-dp9lt" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.424541 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-dp9lt" event={"ID":"4d6d0a8f-2600-47c4-95a0-10f308e493b5","Type":"ContainerDied","Data":"3304e565c579de61731cc043b8b0be89bf7ac35207157a022d9041a6f8f4cb4b"} Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.428586 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-t6bn6" event={"ID":"2abcf43f-6f3b-4804-83a5-ba42fc66f2ec","Type":"ContainerDied","Data":"24bf3019ed6c686bcb675250b277d211fef8c3300ed240bac657ff502d141bda"} Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.428624 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24bf3019ed6c686bcb675250b277d211fef8c3300ed240bac657ff502d141bda" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.428709 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-t6bn6" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.445360 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-874bc5bd9-jqhms" event={"ID":"0c6ef645-08d7-4bd0-aa8a-8181bae77b28","Type":"ContainerDied","Data":"74057c0f494e7b7cf7b23959a450f5cc749a391519fca50368d0a9113e2d7193"} Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.445503 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-874bc5bd9-jqhms" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.500940 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-dp9lt"] Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.510192 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-dp9lt"] Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.528180 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lrm4d"] Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.540046 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lrm4d"] Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.556033 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-874bc5bd9-jqhms"] Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.562623 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-874bc5bd9-jqhms"] Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.850979 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c6ef645-08d7-4bd0-aa8a-8181bae77b28" path="/var/lib/kubelet/pods/0c6ef645-08d7-4bd0-aa8a-8181bae77b28/volumes" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.851704 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d6d0a8f-2600-47c4-95a0-10f308e493b5" path="/var/lib/kubelet/pods/4d6d0a8f-2600-47c4-95a0-10f308e493b5/volumes" Feb 14 11:27:05 crc kubenswrapper[4904]: I0214 11:27:05.852467 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2bcaf28-8b82-460c-9535-b9404a3d19a9" path="/var/lib/kubelet/pods/b2bcaf28-8b82-460c-9535-b9404a3d19a9/volumes" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.063946 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-zqqbz"] Feb 14 11:27:06 crc kubenswrapper[4904]: E0214 11:27:06.072889 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2bcaf28-8b82-460c-9535-b9404a3d19a9" containerName="extract-content" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.072906 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2bcaf28-8b82-460c-9535-b9404a3d19a9" containerName="extract-content" Feb 14 11:27:06 crc kubenswrapper[4904]: E0214 11:27:06.072920 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d6d0a8f-2600-47c4-95a0-10f308e493b5" containerName="dnsmasq-dns" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.072926 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d6d0a8f-2600-47c4-95a0-10f308e493b5" containerName="dnsmasq-dns" Feb 14 11:27:06 crc kubenswrapper[4904]: E0214 11:27:06.072938 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2abcf43f-6f3b-4804-83a5-ba42fc66f2ec" containerName="neutron-db-sync" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.072945 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="2abcf43f-6f3b-4804-83a5-ba42fc66f2ec" containerName="neutron-db-sync" Feb 14 11:27:06 crc kubenswrapper[4904]: E0214 11:27:06.072956 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2bcaf28-8b82-460c-9535-b9404a3d19a9" containerName="registry-server" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.072962 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2bcaf28-8b82-460c-9535-b9404a3d19a9" containerName="registry-server" Feb 14 11:27:06 crc kubenswrapper[4904]: E0214 11:27:06.072981 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d6d0a8f-2600-47c4-95a0-10f308e493b5" containerName="init" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.072987 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d6d0a8f-2600-47c4-95a0-10f308e493b5" containerName="init" Feb 14 11:27:06 crc kubenswrapper[4904]: E0214 11:27:06.072999 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2bcaf28-8b82-460c-9535-b9404a3d19a9" containerName="extract-utilities" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.073005 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2bcaf28-8b82-460c-9535-b9404a3d19a9" containerName="extract-utilities" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.073165 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2bcaf28-8b82-460c-9535-b9404a3d19a9" containerName="registry-server" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.073173 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d6d0a8f-2600-47c4-95a0-10f308e493b5" containerName="dnsmasq-dns" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.073187 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="2abcf43f-6f3b-4804-83a5-ba42fc66f2ec" containerName="neutron-db-sync" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.074052 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.099016 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-zqqbz"] Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.169341 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-55f65b5df6-s2ts6"] Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.189114 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.197780 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.197990 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.202274 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-rjj2h" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.202596 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.240230 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-55f65b5df6-s2ts6"] Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.253087 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-dns-svc\") pod \"dnsmasq-dns-55f844cf75-zqqbz\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.253165 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-zqqbz\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.253187 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-zqqbz\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.253251 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-zqqbz\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.253327 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-config\") pod \"dnsmasq-dns-55f844cf75-zqqbz\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.253365 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjmgp\" (UniqueName: \"kubernetes.io/projected/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-kube-api-access-rjmgp\") pod \"dnsmasq-dns-55f844cf75-zqqbz\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.355034 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-config\") pod \"dnsmasq-dns-55f844cf75-zqqbz\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.355326 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-httpd-config\") pod \"neutron-55f65b5df6-s2ts6\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.355352 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjmgp\" (UniqueName: \"kubernetes.io/projected/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-kube-api-access-rjmgp\") pod \"dnsmasq-dns-55f844cf75-zqqbz\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.355404 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-combined-ca-bundle\") pod \"neutron-55f65b5df6-s2ts6\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.355441 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-ovndb-tls-certs\") pod \"neutron-55f65b5df6-s2ts6\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.355464 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-dns-svc\") pod \"dnsmasq-dns-55f844cf75-zqqbz\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.355492 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-zqqbz\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.355509 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-zqqbz\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.355529 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-config\") pod \"neutron-55f65b5df6-s2ts6\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.355562 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-zqqbz\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.355596 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkr6t\" (UniqueName: \"kubernetes.io/projected/6db2b9f3-7f93-47e8-b493-eaed893c2b50-kube-api-access-nkr6t\") pod \"neutron-55f65b5df6-s2ts6\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.356495 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-config\") pod \"dnsmasq-dns-55f844cf75-zqqbz\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.357455 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-dns-svc\") pod \"dnsmasq-dns-55f844cf75-zqqbz\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.358962 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-zqqbz\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.359677 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-zqqbz\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.366233 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-zqqbz\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.412920 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjmgp\" (UniqueName: \"kubernetes.io/projected/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-kube-api-access-rjmgp\") pod \"dnsmasq-dns-55f844cf75-zqqbz\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: E0214 11:27:06.415065 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Feb 14 11:27:06 crc kubenswrapper[4904]: E0214 11:27:06.415198 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bf8c7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-nnlq9_openstack(c698a36a-d049-4fdb-b279-1b62d1f807d3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 14 11:27:06 crc kubenswrapper[4904]: E0214 11:27:06.416375 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-nnlq9" podUID="c698a36a-d049-4fdb-b279-1b62d1f807d3" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.457138 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-config\") pod \"neutron-55f65b5df6-s2ts6\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.457236 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkr6t\" (UniqueName: \"kubernetes.io/projected/6db2b9f3-7f93-47e8-b493-eaed893c2b50-kube-api-access-nkr6t\") pod \"neutron-55f65b5df6-s2ts6\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.457283 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-httpd-config\") pod \"neutron-55f65b5df6-s2ts6\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.457346 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-combined-ca-bundle\") pod \"neutron-55f65b5df6-s2ts6\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.457385 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-ovndb-tls-certs\") pod \"neutron-55f65b5df6-s2ts6\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.464298 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-config\") pod \"neutron-55f65b5df6-s2ts6\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.465748 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-ovndb-tls-certs\") pod \"neutron-55f65b5df6-s2ts6\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.466803 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-httpd-config\") pod \"neutron-55f65b5df6-s2ts6\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.468229 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-combined-ca-bundle\") pod \"neutron-55f65b5df6-s2ts6\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.486689 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkr6t\" (UniqueName: \"kubernetes.io/projected/6db2b9f3-7f93-47e8-b493-eaed893c2b50-kube-api-access-nkr6t\") pod \"neutron-55f65b5df6-s2ts6\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.502008 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8v2xr" event={"ID":"a9fd9edc-0058-417c-baa6-9fbff5035af5","Type":"ContainerDied","Data":"ebd6623dce31d58bfde727984093e1af8c772c41118a2c8aa7c4d42ef1062443"} Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.502057 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebd6623dce31d58bfde727984093e1af8c772c41118a2c8aa7c4d42ef1062443" Feb 14 11:27:06 crc kubenswrapper[4904]: E0214 11:27:06.505727 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-nnlq9" podUID="c698a36a-d049-4fdb-b279-1b62d1f807d3" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.515095 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.523079 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8v2xr" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.667602 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9fd9edc-0058-417c-baa6-9fbff5035af5-catalog-content\") pod \"a9fd9edc-0058-417c-baa6-9fbff5035af5\" (UID: \"a9fd9edc-0058-417c-baa6-9fbff5035af5\") " Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.667710 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9fd9edc-0058-417c-baa6-9fbff5035af5-utilities\") pod \"a9fd9edc-0058-417c-baa6-9fbff5035af5\" (UID: \"a9fd9edc-0058-417c-baa6-9fbff5035af5\") " Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.667763 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw6jj\" (UniqueName: \"kubernetes.io/projected/a9fd9edc-0058-417c-baa6-9fbff5035af5-kube-api-access-qw6jj\") pod \"a9fd9edc-0058-417c-baa6-9fbff5035af5\" (UID: \"a9fd9edc-0058-417c-baa6-9fbff5035af5\") " Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.670000 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9fd9edc-0058-417c-baa6-9fbff5035af5-utilities" (OuterVolumeSpecName: "utilities") pod "a9fd9edc-0058-417c-baa6-9fbff5035af5" (UID: "a9fd9edc-0058-417c-baa6-9fbff5035af5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.680032 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9fd9edc-0058-417c-baa6-9fbff5035af5-kube-api-access-qw6jj" (OuterVolumeSpecName: "kube-api-access-qw6jj") pod "a9fd9edc-0058-417c-baa6-9fbff5035af5" (UID: "a9fd9edc-0058-417c-baa6-9fbff5035af5"). InnerVolumeSpecName "kube-api-access-qw6jj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.694518 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.771011 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9fd9edc-0058-417c-baa6-9fbff5035af5-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.771048 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw6jj\" (UniqueName: \"kubernetes.io/projected/a9fd9edc-0058-417c-baa6-9fbff5035af5-kube-api-access-qw6jj\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.834370 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9fd9edc-0058-417c-baa6-9fbff5035af5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a9fd9edc-0058-417c-baa6-9fbff5035af5" (UID: "a9fd9edc-0058-417c-baa6-9fbff5035af5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:27:06 crc kubenswrapper[4904]: I0214 11:27:06.873192 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9fd9edc-0058-417c-baa6-9fbff5035af5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:07 crc kubenswrapper[4904]: I0214 11:27:07.175309 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:27:07 crc kubenswrapper[4904]: I0214 11:27:07.235640 4904 scope.go:117] "RemoveContainer" containerID="fec6c3df38894bcf3f8752aac31551259a23a5a3fb9b2c2918e0b43dc674e99a" Feb 14 11:27:07 crc kubenswrapper[4904]: W0214 11:27:07.240783 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a959004_612f_45ee_82ee_9d1fe7af7eba.slice/crio-9be91377f1b8adcf629d262e8514611f3212a26d790eb0298ad65b7b91c26a95 WatchSource:0}: Error finding container 9be91377f1b8adcf629d262e8514611f3212a26d790eb0298ad65b7b91c26a95: Status 404 returned error can't find the container with id 9be91377f1b8adcf629d262e8514611f3212a26d790eb0298ad65b7b91c26a95 Feb 14 11:27:07 crc kubenswrapper[4904]: I0214 11:27:07.467304 4904 scope.go:117] "RemoveContainer" containerID="f8b1e39b3e8f22c8cbe144a51482aaf198a4fbf2c54ceadf14eaeae193414c3e" Feb 14 11:27:07 crc kubenswrapper[4904]: I0214 11:27:07.529802 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2a959004-612f-45ee-82ee-9d1fe7af7eba","Type":"ContainerStarted","Data":"9be91377f1b8adcf629d262e8514611f3212a26d790eb0298ad65b7b91c26a95"} Feb 14 11:27:07 crc kubenswrapper[4904]: I0214 11:27:07.532025 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8v2xr" Feb 14 11:27:07 crc kubenswrapper[4904]: I0214 11:27:07.567286 4904 scope.go:117] "RemoveContainer" containerID="68bcf29edaf47bf7889aa5762ca5390fc72fa16206dd5f0776a150a1958a2278" Feb 14 11:27:07 crc kubenswrapper[4904]: I0214 11:27:07.756488 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-dp9lt" podUID="4d6d0a8f-2600-47c4-95a0-10f308e493b5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.109:5353: i/o timeout" Feb 14 11:27:07 crc kubenswrapper[4904]: I0214 11:27:07.839285 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8v2xr"] Feb 14 11:27:07 crc kubenswrapper[4904]: I0214 11:27:07.894023 4904 scope.go:117] "RemoveContainer" containerID="5a3f35624c4a676928fb933326f2b0137135f8c9e439ee555d4fcd7f836c9026" Feb 14 11:27:07 crc kubenswrapper[4904]: I0214 11:27:07.959470 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8v2xr"] Feb 14 11:27:07 crc kubenswrapper[4904]: I0214 11:27:07.959514 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-v6zmt"] Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.099081 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.157178 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-zqqbz"] Feb 14 11:27:08 crc kubenswrapper[4904]: W0214 11:27:08.179793 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5eaaa567_fbe1_4db5_822a_8b6f1ff773aa.slice/crio-fc9619619b133fd11c3cf64650614957dcf50fa8a8e414f35f7fbec0544e42c8 WatchSource:0}: Error finding container fc9619619b133fd11c3cf64650614957dcf50fa8a8e414f35f7fbec0544e42c8: Status 404 returned error can't find the container with id fc9619619b133fd11c3cf64650614957dcf50fa8a8e414f35f7fbec0544e42c8 Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.280698 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-55f65b5df6-s2ts6"] Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.548427 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16339330-a7b5-471a-ad7e-1a8d4627c3e2","Type":"ContainerStarted","Data":"c9dc77b78a504691a7a3af90a1868fc0b6ba517a927791a09b98aab71a44ff6d"} Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.552761 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ft76h" event={"ID":"60ac5de4-7c8e-4065-a846-3de7a3cffcc0","Type":"ContainerStarted","Data":"b37fbd340585a792ef4d238687d76139e067db8414b45090fb4d705a9b528d75"} Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.558403 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"068d8b3e-9bde-4ac2-bcab-1c3e46670bba","Type":"ContainerStarted","Data":"3eb7ab2eba0e676c34a243469aff7c63fb338818f5ff6e2d94be24861e684ac1"} Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.562298 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" event={"ID":"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa","Type":"ContainerStarted","Data":"fc9619619b133fd11c3cf64650614957dcf50fa8a8e414f35f7fbec0544e42c8"} Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.571208 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v6zmt" event={"ID":"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1","Type":"ContainerStarted","Data":"2f654fcc93b3f55b28a5fe2cb572abfbf0d7e75bf3484ef586fdda4fcdd6723b"} Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.571253 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v6zmt" event={"ID":"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1","Type":"ContainerStarted","Data":"3c71ec1318ead0684489648cf1ed5487a87ba64fa72433d7aa8d12729f041731"} Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.578999 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-687657b496-tz596" event={"ID":"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393","Type":"ContainerStarted","Data":"582bc45e7c749f481aab958e03a53b45ff101180b75c9d8a70ed2568b4a2948b"} Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.583743 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-ft76h" podStartSLOduration=6.644141318 podStartE2EDuration="39.58372716s" podCreationTimestamp="2026-02-14 11:26:29 +0000 UTC" firstStartedPulling="2026-02-14 11:26:31.623926391 +0000 UTC m=+982.436691052" lastFinishedPulling="2026-02-14 11:27:04.563512233 +0000 UTC m=+1015.376276894" observedRunningTime="2026-02-14 11:27:08.582554748 +0000 UTC m=+1019.395319409" watchObservedRunningTime="2026-02-14 11:27:08.58372716 +0000 UTC m=+1019.396491811" Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.600673 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55f65b5df6-s2ts6" event={"ID":"6db2b9f3-7f93-47e8-b493-eaed893c2b50","Type":"ContainerStarted","Data":"ff0ca90250e14827d90c10778ad1233db548d045ac850cdc51c3071b180d1f67"} Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.608613 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bc9d549b8-hdn6n" event={"ID":"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71","Type":"ContainerStarted","Data":"3877326e39865783a2a0f4296a7397360e5194ffdc9f51979daccec89f965141"} Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.762899 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-v6zmt" podStartSLOduration=12.762880802 podStartE2EDuration="12.762880802s" podCreationTimestamp="2026-02-14 11:26:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:27:08.606039712 +0000 UTC m=+1019.418804373" watchObservedRunningTime="2026-02-14 11:27:08.762880802 +0000 UTC m=+1019.575645463" Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.805506 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-58c9948449-g7lbx"] Feb 14 11:27:08 crc kubenswrapper[4904]: E0214 11:27:08.805937 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9fd9edc-0058-417c-baa6-9fbff5035af5" containerName="registry-server" Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.805955 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9fd9edc-0058-417c-baa6-9fbff5035af5" containerName="registry-server" Feb 14 11:27:08 crc kubenswrapper[4904]: E0214 11:27:08.805974 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9fd9edc-0058-417c-baa6-9fbff5035af5" containerName="extract-content" Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.805982 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9fd9edc-0058-417c-baa6-9fbff5035af5" containerName="extract-content" Feb 14 11:27:08 crc kubenswrapper[4904]: E0214 11:27:08.806024 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9fd9edc-0058-417c-baa6-9fbff5035af5" containerName="extract-utilities" Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.806032 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9fd9edc-0058-417c-baa6-9fbff5035af5" containerName="extract-utilities" Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.806261 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9fd9edc-0058-417c-baa6-9fbff5035af5" containerName="registry-server" Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.807410 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.811189 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.811360 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.868726 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-58c9948449-g7lbx"] Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.954959 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-ovndb-tls-certs\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.955016 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-internal-tls-certs\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.955082 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-combined-ca-bundle\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.955126 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-public-tls-certs\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.955204 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-httpd-config\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.955236 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnskp\" (UniqueName: \"kubernetes.io/projected/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-kube-api-access-mnskp\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:08 crc kubenswrapper[4904]: I0214 11:27:08.955270 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-config\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.057243 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-internal-tls-certs\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.057318 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-combined-ca-bundle\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.057356 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-public-tls-certs\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.057426 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-httpd-config\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.057445 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnskp\" (UniqueName: \"kubernetes.io/projected/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-kube-api-access-mnskp\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.057469 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-config\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.057512 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-ovndb-tls-certs\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.070063 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-public-tls-certs\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.072045 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-ovndb-tls-certs\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.075459 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-config\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.076487 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-internal-tls-certs\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.077104 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-combined-ca-bundle\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.082228 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-httpd-config\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.110420 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnskp\" (UniqueName: \"kubernetes.io/projected/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-kube-api-access-mnskp\") pod \"neutron-58c9948449-g7lbx\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.239248 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.637711 4904 generic.go:334] "Generic (PLEG): container finished" podID="5eaaa567-fbe1-4db5-822a-8b6f1ff773aa" containerID="7d5e8e8081b7b6b9665f4e4356f8c5968e9dceee7028186d474ec75aa35a72a8" exitCode=0 Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.638293 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" event={"ID":"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa","Type":"ContainerDied","Data":"7d5e8e8081b7b6b9665f4e4356f8c5968e9dceee7028186d474ec75aa35a72a8"} Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.677017 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2a959004-612f-45ee-82ee-9d1fe7af7eba","Type":"ContainerStarted","Data":"60066d7089ba0c88d30e7c41fb469f1d08af664bc554bff90b08522c23dcb662"} Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.719804 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bc9d549b8-hdn6n" event={"ID":"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71","Type":"ContainerStarted","Data":"14da3471fa5fd4d6ec42e957717617fb2dea34019a855f12f66753cef241842c"} Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.737587 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-687657b496-tz596" event={"ID":"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393","Type":"ContainerStarted","Data":"609e50be7f47738cb94497fefee796e278d503fdc50869f2e264123ccbb09fa6"} Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.786976 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55f65b5df6-s2ts6" event={"ID":"6db2b9f3-7f93-47e8-b493-eaed893c2b50","Type":"ContainerStarted","Data":"ae3fafc76a52c56cbeeeab7e7d6b63689bc9a8947fc5707d06ef01e25c10fc50"} Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.824386 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6bc9d549b8-hdn6n" podStartSLOduration=16.833471363 podStartE2EDuration="31.824367576s" podCreationTimestamp="2026-02-14 11:26:38 +0000 UTC" firstStartedPulling="2026-02-14 11:26:52.288626039 +0000 UTC m=+1003.101390710" lastFinishedPulling="2026-02-14 11:27:07.279522262 +0000 UTC m=+1018.092286923" observedRunningTime="2026-02-14 11:27:09.819818131 +0000 UTC m=+1020.632582792" watchObservedRunningTime="2026-02-14 11:27:09.824367576 +0000 UTC m=+1020.637132237" Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.904459 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-687657b496-tz596" podStartSLOduration=7.63763382 podStartE2EDuration="31.904438501s" podCreationTimestamp="2026-02-14 11:26:38 +0000 UTC" firstStartedPulling="2026-02-14 11:26:43.013987935 +0000 UTC m=+993.826752596" lastFinishedPulling="2026-02-14 11:27:07.280792616 +0000 UTC m=+1018.093557277" observedRunningTime="2026-02-14 11:27:09.872423264 +0000 UTC m=+1020.685187915" watchObservedRunningTime="2026-02-14 11:27:09.904438501 +0000 UTC m=+1020.717203162" Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.964052 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9fd9edc-0058-417c-baa6-9fbff5035af5" path="/var/lib/kubelet/pods/a9fd9edc-0058-417c-baa6-9fbff5035af5/volumes" Feb 14 11:27:09 crc kubenswrapper[4904]: I0214 11:27:09.964973 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"068d8b3e-9bde-4ac2-bcab-1c3e46670bba","Type":"ContainerStarted","Data":"3cb6e0644c18f4726141d54faf948ed7ccebf0a4e5ee9165ae1ba3faa19dd21b"} Feb 14 11:27:10 crc kubenswrapper[4904]: I0214 11:27:10.177520 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-58c9948449-g7lbx"] Feb 14 11:27:10 crc kubenswrapper[4904]: W0214 11:27:10.232570 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4cc31d3d_1f06_48a1_9856_e1a997bfb11e.slice/crio-abc724f0405ac23e7f36a76f074c375906ad3ae83471a52035a6ca5949bbeabb WatchSource:0}: Error finding container abc724f0405ac23e7f36a76f074c375906ad3ae83471a52035a6ca5949bbeabb: Status 404 returned error can't find the container with id abc724f0405ac23e7f36a76f074c375906ad3ae83471a52035a6ca5949bbeabb Feb 14 11:27:10 crc kubenswrapper[4904]: I0214 11:27:10.941325 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55f65b5df6-s2ts6" event={"ID":"6db2b9f3-7f93-47e8-b493-eaed893c2b50","Type":"ContainerStarted","Data":"bd9338ff572e4987dc33310a36fef6ebc04cc99afbddefc451c4a431ad71fa42"} Feb 14 11:27:10 crc kubenswrapper[4904]: I0214 11:27:10.942940 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:27:10 crc kubenswrapper[4904]: I0214 11:27:10.947645 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58c9948449-g7lbx" event={"ID":"4cc31d3d-1f06-48a1-9856-e1a997bfb11e","Type":"ContainerStarted","Data":"df39d52890d4f62c453dfa0452f433a477f0e1ede96fc0bb86defbf9ea1c50a1"} Feb 14 11:27:10 crc kubenswrapper[4904]: I0214 11:27:10.947671 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58c9948449-g7lbx" event={"ID":"4cc31d3d-1f06-48a1-9856-e1a997bfb11e","Type":"ContainerStarted","Data":"abc724f0405ac23e7f36a76f074c375906ad3ae83471a52035a6ca5949bbeabb"} Feb 14 11:27:10 crc kubenswrapper[4904]: I0214 11:27:10.950207 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"068d8b3e-9bde-4ac2-bcab-1c3e46670bba","Type":"ContainerStarted","Data":"1622edf726b9b4d539db9c2f4fa2c5488024e9c391291015fa1f502a76ef8a9b"} Feb 14 11:27:10 crc kubenswrapper[4904]: I0214 11:27:10.954212 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" event={"ID":"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa","Type":"ContainerStarted","Data":"5478a0462c7ac487dd027e1b03f15769668ff98a095725635740e868c8f50bd1"} Feb 14 11:27:10 crc kubenswrapper[4904]: I0214 11:27:10.954651 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:10 crc kubenswrapper[4904]: I0214 11:27:10.960229 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2a959004-612f-45ee-82ee-9d1fe7af7eba","Type":"ContainerStarted","Data":"7233d6630e459d23bd14e59aaaf7361d14eddf1e0c57aa2a7a93459663d61fbd"} Feb 14 11:27:10 crc kubenswrapper[4904]: I0214 11:27:10.983790 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-55f65b5df6-s2ts6" podStartSLOduration=4.983771535 podStartE2EDuration="4.983771535s" podCreationTimestamp="2026-02-14 11:27:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:27:10.971813527 +0000 UTC m=+1021.784578188" watchObservedRunningTime="2026-02-14 11:27:10.983771535 +0000 UTC m=+1021.796536196" Feb 14 11:27:11 crc kubenswrapper[4904]: I0214 11:27:11.014166 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=31.014151268 podStartE2EDuration="31.014151268s" podCreationTimestamp="2026-02-14 11:26:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:27:10.997517992 +0000 UTC m=+1021.810282653" watchObservedRunningTime="2026-02-14 11:27:11.014151268 +0000 UTC m=+1021.826915929" Feb 14 11:27:11 crc kubenswrapper[4904]: I0214 11:27:11.027564 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=31.027551366 podStartE2EDuration="31.027551366s" podCreationTimestamp="2026-02-14 11:26:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:27:11.0251758 +0000 UTC m=+1021.837940461" watchObservedRunningTime="2026-02-14 11:27:11.027551366 +0000 UTC m=+1021.840316027" Feb 14 11:27:11 crc kubenswrapper[4904]: I0214 11:27:11.067903 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" podStartSLOduration=5.06787149 podStartE2EDuration="5.06787149s" podCreationTimestamp="2026-02-14 11:27:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:27:11.050819973 +0000 UTC m=+1021.863584654" watchObservedRunningTime="2026-02-14 11:27:11.06787149 +0000 UTC m=+1021.880636171" Feb 14 11:27:11 crc kubenswrapper[4904]: I0214 11:27:11.975112 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58c9948449-g7lbx" event={"ID":"4cc31d3d-1f06-48a1-9856-e1a997bfb11e","Type":"ContainerStarted","Data":"eea579d5897474d21fdec00430fbfcb39d8f2d878d90ca0eb9305bd3657fc893"} Feb 14 11:27:11 crc kubenswrapper[4904]: I0214 11:27:11.999367 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-58c9948449-g7lbx" podStartSLOduration=3.99934923 podStartE2EDuration="3.99934923s" podCreationTimestamp="2026-02-14 11:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:27:11.997439758 +0000 UTC m=+1022.810204429" watchObservedRunningTime="2026-02-14 11:27:11.99934923 +0000 UTC m=+1022.812113891" Feb 14 11:27:12 crc kubenswrapper[4904]: I0214 11:27:12.989210 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:13 crc kubenswrapper[4904]: I0214 11:27:13.999054 4904 generic.go:334] "Generic (PLEG): container finished" podID="60ac5de4-7c8e-4065-a846-3de7a3cffcc0" containerID="b37fbd340585a792ef4d238687d76139e067db8414b45090fb4d705a9b528d75" exitCode=0 Feb 14 11:27:13 crc kubenswrapper[4904]: I0214 11:27:13.999138 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ft76h" event={"ID":"60ac5de4-7c8e-4065-a846-3de7a3cffcc0","Type":"ContainerDied","Data":"b37fbd340585a792ef4d238687d76139e067db8414b45090fb4d705a9b528d75"} Feb 14 11:27:16 crc kubenswrapper[4904]: I0214 11:27:16.023545 4904 generic.go:334] "Generic (PLEG): container finished" podID="ec37844d-1c49-4aa6-b9e1-87ba60eadbe1" containerID="2f654fcc93b3f55b28a5fe2cb572abfbf0d7e75bf3484ef586fdda4fcdd6723b" exitCode=0 Feb 14 11:27:16 crc kubenswrapper[4904]: I0214 11:27:16.023630 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v6zmt" event={"ID":"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1","Type":"ContainerDied","Data":"2f654fcc93b3f55b28a5fe2cb572abfbf0d7e75bf3484ef586fdda4fcdd6723b"} Feb 14 11:27:16 crc kubenswrapper[4904]: I0214 11:27:16.695978 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:16 crc kubenswrapper[4904]: I0214 11:27:16.749318 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-5wt4c"] Feb 14 11:27:16 crc kubenswrapper[4904]: I0214 11:27:16.749522 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" podUID="df0d5e85-642d-4fab-a77d-337c4a4518d6" containerName="dnsmasq-dns" containerID="cri-o://613f7e92e44f4c81d170f628b5fcb673319c5698a425bc631d940c9ecbff4c13" gracePeriod=10 Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.062577 4904 generic.go:334] "Generic (PLEG): container finished" podID="df0d5e85-642d-4fab-a77d-337c4a4518d6" containerID="613f7e92e44f4c81d170f628b5fcb673319c5698a425bc631d940c9ecbff4c13" exitCode=0 Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.062657 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" event={"ID":"df0d5e85-642d-4fab-a77d-337c4a4518d6","Type":"ContainerDied","Data":"613f7e92e44f4c81d170f628b5fcb673319c5698a425bc631d940c9ecbff4c13"} Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.365799 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ft76h" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.502872 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-config-data\") pod \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.503005 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-logs\") pod \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.503041 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nft6t\" (UniqueName: \"kubernetes.io/projected/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-kube-api-access-nft6t\") pod \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.503083 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-combined-ca-bundle\") pod \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.503108 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-scripts\") pod \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\" (UID: \"60ac5de4-7c8e-4065-a846-3de7a3cffcc0\") " Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.504453 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-logs" (OuterVolumeSpecName: "logs") pod "60ac5de4-7c8e-4065-a846-3de7a3cffcc0" (UID: "60ac5de4-7c8e-4065-a846-3de7a3cffcc0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.516046 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-scripts" (OuterVolumeSpecName: "scripts") pod "60ac5de4-7c8e-4065-a846-3de7a3cffcc0" (UID: "60ac5de4-7c8e-4065-a846-3de7a3cffcc0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.522687 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-kube-api-access-nft6t" (OuterVolumeSpecName: "kube-api-access-nft6t") pod "60ac5de4-7c8e-4065-a846-3de7a3cffcc0" (UID: "60ac5de4-7c8e-4065-a846-3de7a3cffcc0"). InnerVolumeSpecName "kube-api-access-nft6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.605649 4904 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-logs\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.605695 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nft6t\" (UniqueName: \"kubernetes.io/projected/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-kube-api-access-nft6t\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.605710 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.610231 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60ac5de4-7c8e-4065-a846-3de7a3cffcc0" (UID: "60ac5de4-7c8e-4065-a846-3de7a3cffcc0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.612159 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.612354 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-config-data" (OuterVolumeSpecName: "config-data") pod "60ac5de4-7c8e-4065-a846-3de7a3cffcc0" (UID: "60ac5de4-7c8e-4065-a846-3de7a3cffcc0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.709310 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-config-data\") pod \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.709351 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-fernet-keys\") pod \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.709426 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-combined-ca-bundle\") pod \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.709462 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-scripts\") pod \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.709520 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-credential-keys\") pod \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.709543 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2j7s\" (UniqueName: \"kubernetes.io/projected/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-kube-api-access-h2j7s\") pod \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\" (UID: \"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1\") " Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.709984 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.710001 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60ac5de4-7c8e-4065-a846-3de7a3cffcc0-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.732335 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ec37844d-1c49-4aa6-b9e1-87ba60eadbe1" (UID: "ec37844d-1c49-4aa6-b9e1-87ba60eadbe1"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.735039 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-kube-api-access-h2j7s" (OuterVolumeSpecName: "kube-api-access-h2j7s") pod "ec37844d-1c49-4aa6-b9e1-87ba60eadbe1" (UID: "ec37844d-1c49-4aa6-b9e1-87ba60eadbe1"). InnerVolumeSpecName "kube-api-access-h2j7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.735096 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ec37844d-1c49-4aa6-b9e1-87ba60eadbe1" (UID: "ec37844d-1c49-4aa6-b9e1-87ba60eadbe1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.761030 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-scripts" (OuterVolumeSpecName: "scripts") pod "ec37844d-1c49-4aa6-b9e1-87ba60eadbe1" (UID: "ec37844d-1c49-4aa6-b9e1-87ba60eadbe1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.788966 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-config-data" (OuterVolumeSpecName: "config-data") pod "ec37844d-1c49-4aa6-b9e1-87ba60eadbe1" (UID: "ec37844d-1c49-4aa6-b9e1-87ba60eadbe1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.804380 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec37844d-1c49-4aa6-b9e1-87ba60eadbe1" (UID: "ec37844d-1c49-4aa6-b9e1-87ba60eadbe1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.811935 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.811969 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.811979 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2j7s\" (UniqueName: \"kubernetes.io/projected/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-kube-api-access-h2j7s\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.811991 4904 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.812000 4904 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.812008 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:17 crc kubenswrapper[4904]: I0214 11:27:17.879476 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.025346 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-dns-svc\") pod \"df0d5e85-642d-4fab-a77d-337c4a4518d6\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.025389 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-ovsdbserver-sb\") pod \"df0d5e85-642d-4fab-a77d-337c4a4518d6\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.025448 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-config\") pod \"df0d5e85-642d-4fab-a77d-337c4a4518d6\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.025573 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-dns-swift-storage-0\") pod \"df0d5e85-642d-4fab-a77d-337c4a4518d6\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.025591 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpwqj\" (UniqueName: \"kubernetes.io/projected/df0d5e85-642d-4fab-a77d-337c4a4518d6-kube-api-access-wpwqj\") pod \"df0d5e85-642d-4fab-a77d-337c4a4518d6\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.025612 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-ovsdbserver-nb\") pod \"df0d5e85-642d-4fab-a77d-337c4a4518d6\" (UID: \"df0d5e85-642d-4fab-a77d-337c4a4518d6\") " Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.047181 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df0d5e85-642d-4fab-a77d-337c4a4518d6-kube-api-access-wpwqj" (OuterVolumeSpecName: "kube-api-access-wpwqj") pod "df0d5e85-642d-4fab-a77d-337c4a4518d6" (UID: "df0d5e85-642d-4fab-a77d-337c4a4518d6"). InnerVolumeSpecName "kube-api-access-wpwqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.076608 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16339330-a7b5-471a-ad7e-1a8d4627c3e2","Type":"ContainerStarted","Data":"7698083f123e53016f0ffaa8355298315cb001448a6b1604956e2988bba73597"} Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.078276 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ft76h" event={"ID":"60ac5de4-7c8e-4065-a846-3de7a3cffcc0","Type":"ContainerDied","Data":"541a3dfa767b27a430b04c055cfc2580684232830d887cddbc03fc91ba12fdc9"} Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.078294 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="541a3dfa767b27a430b04c055cfc2580684232830d887cddbc03fc91ba12fdc9" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.078354 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ft76h" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.081223 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" event={"ID":"df0d5e85-642d-4fab-a77d-337c4a4518d6","Type":"ContainerDied","Data":"39b62eecc442c6db7887ed31c8ff69711b14ec1aaf3b11345182a53fe3ce70df"} Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.081246 4904 scope.go:117] "RemoveContainer" containerID="613f7e92e44f4c81d170f628b5fcb673319c5698a425bc631d940c9ecbff4c13" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.081336 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-5wt4c" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.091982 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sflms" event={"ID":"6b349fb1-5e71-45ff-88e9-03592bfe2504","Type":"ContainerStarted","Data":"752aa5a87f01259b48e8dea972209d56a5ec87f518b1e02f7097ac9ce281ea76"} Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.107515 4904 scope.go:117] "RemoveContainer" containerID="46d5f87b0312e65ebe0479ad137dabccf046ed0c414e4b6ba23423036fcc0147" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.109510 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v6zmt" event={"ID":"ec37844d-1c49-4aa6-b9e1-87ba60eadbe1","Type":"ContainerDied","Data":"3c71ec1318ead0684489648cf1ed5487a87ba64fa72433d7aa8d12729f041731"} Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.109554 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c71ec1318ead0684489648cf1ed5487a87ba64fa72433d7aa8d12729f041731" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.109638 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v6zmt" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.126970 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "df0d5e85-642d-4fab-a77d-337c4a4518d6" (UID: "df0d5e85-642d-4fab-a77d-337c4a4518d6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.128175 4904 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.128201 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpwqj\" (UniqueName: \"kubernetes.io/projected/df0d5e85-642d-4fab-a77d-337c4a4518d6-kube-api-access-wpwqj\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.138469 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-config" (OuterVolumeSpecName: "config") pod "df0d5e85-642d-4fab-a77d-337c4a4518d6" (UID: "df0d5e85-642d-4fab-a77d-337c4a4518d6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.141138 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "df0d5e85-642d-4fab-a77d-337c4a4518d6" (UID: "df0d5e85-642d-4fab-a77d-337c4a4518d6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.162615 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "df0d5e85-642d-4fab-a77d-337c4a4518d6" (UID: "df0d5e85-642d-4fab-a77d-337c4a4518d6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.167192 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-sflms" podStartSLOduration=3.584776065 podStartE2EDuration="49.167178881s" podCreationTimestamp="2026-02-14 11:26:29 +0000 UTC" firstStartedPulling="2026-02-14 11:26:31.681490569 +0000 UTC m=+982.494255230" lastFinishedPulling="2026-02-14 11:27:17.263893385 +0000 UTC m=+1028.076658046" observedRunningTime="2026-02-14 11:27:18.107525445 +0000 UTC m=+1028.920290106" watchObservedRunningTime="2026-02-14 11:27:18.167178881 +0000 UTC m=+1028.979943532" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.170052 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7b79cd46c9-g4bx8"] Feb 14 11:27:18 crc kubenswrapper[4904]: E0214 11:27:18.170424 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df0d5e85-642d-4fab-a77d-337c4a4518d6" containerName="init" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.170440 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="df0d5e85-642d-4fab-a77d-337c4a4518d6" containerName="init" Feb 14 11:27:18 crc kubenswrapper[4904]: E0214 11:27:18.170451 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec37844d-1c49-4aa6-b9e1-87ba60eadbe1" containerName="keystone-bootstrap" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.170457 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec37844d-1c49-4aa6-b9e1-87ba60eadbe1" containerName="keystone-bootstrap" Feb 14 11:27:18 crc kubenswrapper[4904]: E0214 11:27:18.170502 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60ac5de4-7c8e-4065-a846-3de7a3cffcc0" containerName="placement-db-sync" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.170509 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="60ac5de4-7c8e-4065-a846-3de7a3cffcc0" containerName="placement-db-sync" Feb 14 11:27:18 crc kubenswrapper[4904]: E0214 11:27:18.170521 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df0d5e85-642d-4fab-a77d-337c4a4518d6" containerName="dnsmasq-dns" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.170528 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="df0d5e85-642d-4fab-a77d-337c4a4518d6" containerName="dnsmasq-dns" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.170729 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="60ac5de4-7c8e-4065-a846-3de7a3cffcc0" containerName="placement-db-sync" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.170763 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec37844d-1c49-4aa6-b9e1-87ba60eadbe1" containerName="keystone-bootstrap" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.170812 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="df0d5e85-642d-4fab-a77d-337c4a4518d6" containerName="dnsmasq-dns" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.171442 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.174953 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.175144 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.175347 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.175533 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.175630 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.175730 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-57x7t" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.190211 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "df0d5e85-642d-4fab-a77d-337c4a4518d6" (UID: "df0d5e85-642d-4fab-a77d-337c4a4518d6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.191204 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7b79cd46c9-g4bx8"] Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.229721 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-internal-tls-certs\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.230507 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-public-tls-certs\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.230597 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-credential-keys\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.230781 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-config-data\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.230808 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-scripts\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.230880 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-fernet-keys\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.230952 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-combined-ca-bundle\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.231365 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdqsz\" (UniqueName: \"kubernetes.io/projected/ba1e5d40-479d-41ed-953c-3a14733e0673-kube-api-access-pdqsz\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.231638 4904 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.231661 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.231725 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.231737 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df0d5e85-642d-4fab-a77d-337c4a4518d6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.333112 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-public-tls-certs\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.333188 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-credential-keys\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.333239 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-config-data\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.333258 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-scripts\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.333287 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-fernet-keys\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.333318 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-combined-ca-bundle\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.333365 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdqsz\" (UniqueName: \"kubernetes.io/projected/ba1e5d40-479d-41ed-953c-3a14733e0673-kube-api-access-pdqsz\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.333395 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-internal-tls-certs\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.342355 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-public-tls-certs\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.353940 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-config-data\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.354459 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-combined-ca-bundle\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.354767 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-internal-tls-certs\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.355190 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-credential-keys\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.360357 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-scripts\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.361558 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ba1e5d40-479d-41ed-953c-3a14733e0673-fernet-keys\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.363302 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdqsz\" (UniqueName: \"kubernetes.io/projected/ba1e5d40-479d-41ed-953c-3a14733e0673-kube-api-access-pdqsz\") pod \"keystone-7b79cd46c9-g4bx8\" (UID: \"ba1e5d40-479d-41ed-953c-3a14733e0673\") " pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.449896 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-5wt4c"] Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.465821 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-5wt4c"] Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.493648 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.624336 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-bb76f8864-hs52l"] Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.626770 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.628267 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.631078 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.631252 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.631341 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.631623 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-ndvl8" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.660951 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bb76f8864-hs52l"] Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.743298 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dbd404d-9e85-4949-b9a3-194a2708ef89-public-tls-certs\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.743555 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dbd404d-9e85-4949-b9a3-194a2708ef89-logs\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.743579 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2dbd404d-9e85-4949-b9a3-194a2708ef89-scripts\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.743639 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ltsp\" (UniqueName: \"kubernetes.io/projected/2dbd404d-9e85-4949-b9a3-194a2708ef89-kube-api-access-8ltsp\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.743679 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dbd404d-9e85-4949-b9a3-194a2708ef89-config-data\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.743704 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dbd404d-9e85-4949-b9a3-194a2708ef89-internal-tls-certs\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.743761 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dbd404d-9e85-4949-b9a3-194a2708ef89-combined-ca-bundle\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.789282 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.789316 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.792250 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6bc9d549b8-hdn6n" podUID="cfb252c0-ffb2-4d87-9c4e-b56cc2070d71" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.845661 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dbd404d-9e85-4949-b9a3-194a2708ef89-internal-tls-certs\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.845747 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dbd404d-9e85-4949-b9a3-194a2708ef89-combined-ca-bundle\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.845787 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dbd404d-9e85-4949-b9a3-194a2708ef89-public-tls-certs\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.845811 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dbd404d-9e85-4949-b9a3-194a2708ef89-logs\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.845831 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2dbd404d-9e85-4949-b9a3-194a2708ef89-scripts\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.845909 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ltsp\" (UniqueName: \"kubernetes.io/projected/2dbd404d-9e85-4949-b9a3-194a2708ef89-kube-api-access-8ltsp\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.845929 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dbd404d-9e85-4949-b9a3-194a2708ef89-config-data\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.846548 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dbd404d-9e85-4949-b9a3-194a2708ef89-logs\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.854509 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dbd404d-9e85-4949-b9a3-194a2708ef89-internal-tls-certs\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.855823 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dbd404d-9e85-4949-b9a3-194a2708ef89-public-tls-certs\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.860665 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dbd404d-9e85-4949-b9a3-194a2708ef89-combined-ca-bundle\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.860665 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dbd404d-9e85-4949-b9a3-194a2708ef89-config-data\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.861754 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2dbd404d-9e85-4949-b9a3-194a2708ef89-scripts\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.868928 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ltsp\" (UniqueName: \"kubernetes.io/projected/2dbd404d-9e85-4949-b9a3-194a2708ef89-kube-api-access-8ltsp\") pod \"placement-bb76f8864-hs52l\" (UID: \"2dbd404d-9e85-4949-b9a3-194a2708ef89\") " pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.948705 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.949139 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-687657b496-tz596" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.949177 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-687657b496-tz596" Feb 14 11:27:18 crc kubenswrapper[4904]: I0214 11:27:18.951662 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-687657b496-tz596" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 14 11:27:19 crc kubenswrapper[4904]: I0214 11:27:19.211452 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7b79cd46c9-g4bx8"] Feb 14 11:27:19 crc kubenswrapper[4904]: I0214 11:27:19.530590 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bb76f8864-hs52l"] Feb 14 11:27:19 crc kubenswrapper[4904]: I0214 11:27:19.914248 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df0d5e85-642d-4fab-a77d-337c4a4518d6" path="/var/lib/kubelet/pods/df0d5e85-642d-4fab-a77d-337c4a4518d6/volumes" Feb 14 11:27:20 crc kubenswrapper[4904]: I0214 11:27:20.154407 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7b79cd46c9-g4bx8" event={"ID":"ba1e5d40-479d-41ed-953c-3a14733e0673","Type":"ContainerStarted","Data":"74d8765abf8f1196ba3e02b024cf8926ec4634f33337585c93de27c346e55e8a"} Feb 14 11:27:20 crc kubenswrapper[4904]: I0214 11:27:20.154810 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7b79cd46c9-g4bx8" event={"ID":"ba1e5d40-479d-41ed-953c-3a14733e0673","Type":"ContainerStarted","Data":"8686c1fde00377e673555e23a21205f1dbf0dcd553617cfcc2c1ad01a6745fd2"} Feb 14 11:27:20 crc kubenswrapper[4904]: I0214 11:27:20.157106 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:20 crc kubenswrapper[4904]: I0214 11:27:20.174585 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bb76f8864-hs52l" event={"ID":"2dbd404d-9e85-4949-b9a3-194a2708ef89","Type":"ContainerStarted","Data":"d25a81aae02bd6b34801b5d1ef923d3901ef86f9f1bee62dc135fec36b3ec60e"} Feb 14 11:27:20 crc kubenswrapper[4904]: I0214 11:27:20.174623 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bb76f8864-hs52l" event={"ID":"2dbd404d-9e85-4949-b9a3-194a2708ef89","Type":"ContainerStarted","Data":"2224d64a107b32f0a92f267de188681af8070c3f2f1fea0461edaa358aa7f7f0"} Feb 14 11:27:20 crc kubenswrapper[4904]: I0214 11:27:20.188546 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7b79cd46c9-g4bx8" podStartSLOduration=2.188530492 podStartE2EDuration="2.188530492s" podCreationTimestamp="2026-02-14 11:27:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:27:20.179881765 +0000 UTC m=+1030.992646426" watchObservedRunningTime="2026-02-14 11:27:20.188530492 +0000 UTC m=+1031.001295143" Feb 14 11:27:20 crc kubenswrapper[4904]: I0214 11:27:20.641099 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 14 11:27:20 crc kubenswrapper[4904]: I0214 11:27:20.642114 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 14 11:27:20 crc kubenswrapper[4904]: I0214 11:27:20.693645 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 14 11:27:20 crc kubenswrapper[4904]: I0214 11:27:20.739155 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 14 11:27:20 crc kubenswrapper[4904]: I0214 11:27:20.750973 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 14 11:27:20 crc kubenswrapper[4904]: I0214 11:27:20.751447 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 14 11:27:20 crc kubenswrapper[4904]: I0214 11:27:20.816403 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 14 11:27:20 crc kubenswrapper[4904]: I0214 11:27:20.849650 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 14 11:27:21 crc kubenswrapper[4904]: I0214 11:27:21.188543 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nnlq9" event={"ID":"c698a36a-d049-4fdb-b279-1b62d1f807d3","Type":"ContainerStarted","Data":"4d14917974e44826233222d2c4beb085c33e448861d26422e9fcec0bd9877e9a"} Feb 14 11:27:21 crc kubenswrapper[4904]: I0214 11:27:21.194695 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bb76f8864-hs52l" event={"ID":"2dbd404d-9e85-4949-b9a3-194a2708ef89","Type":"ContainerStarted","Data":"05b7614a9ae6da1608d9c3f8a3c990913e677cd0997f065102db04eda9efcf09"} Feb 14 11:27:21 crc kubenswrapper[4904]: I0214 11:27:21.194966 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 14 11:27:21 crc kubenswrapper[4904]: I0214 11:27:21.195051 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 14 11:27:21 crc kubenswrapper[4904]: I0214 11:27:21.196372 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 14 11:27:21 crc kubenswrapper[4904]: I0214 11:27:21.196401 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 14 11:27:21 crc kubenswrapper[4904]: I0214 11:27:21.221571 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-nnlq9" podStartSLOduration=3.873243555 podStartE2EDuration="52.221554146s" podCreationTimestamp="2026-02-14 11:26:29 +0000 UTC" firstStartedPulling="2026-02-14 11:26:31.174850718 +0000 UTC m=+981.987615379" lastFinishedPulling="2026-02-14 11:27:19.523161309 +0000 UTC m=+1030.335925970" observedRunningTime="2026-02-14 11:27:21.211217123 +0000 UTC m=+1032.023981784" watchObservedRunningTime="2026-02-14 11:27:21.221554146 +0000 UTC m=+1032.034318807" Feb 14 11:27:22 crc kubenswrapper[4904]: I0214 11:27:22.206692 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:22 crc kubenswrapper[4904]: I0214 11:27:22.207002 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:23 crc kubenswrapper[4904]: I0214 11:27:23.216018 4904 generic.go:334] "Generic (PLEG): container finished" podID="6b349fb1-5e71-45ff-88e9-03592bfe2504" containerID="752aa5a87f01259b48e8dea972209d56a5ec87f518b1e02f7097ac9ce281ea76" exitCode=0 Feb 14 11:27:23 crc kubenswrapper[4904]: I0214 11:27:23.216103 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sflms" event={"ID":"6b349fb1-5e71-45ff-88e9-03592bfe2504","Type":"ContainerDied","Data":"752aa5a87f01259b48e8dea972209d56a5ec87f518b1e02f7097ac9ce281ea76"} Feb 14 11:27:23 crc kubenswrapper[4904]: I0214 11:27:23.238244 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-bb76f8864-hs52l" podStartSLOduration=5.23822362 podStartE2EDuration="5.23822362s" podCreationTimestamp="2026-02-14 11:27:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:27:21.252761502 +0000 UTC m=+1032.065526163" watchObservedRunningTime="2026-02-14 11:27:23.23822362 +0000 UTC m=+1034.050988281" Feb 14 11:27:25 crc kubenswrapper[4904]: I0214 11:27:25.685344 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 14 11:27:25 crc kubenswrapper[4904]: I0214 11:27:25.685848 4904 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 14 11:27:25 crc kubenswrapper[4904]: I0214 11:27:25.712321 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 14 11:27:25 crc kubenswrapper[4904]: I0214 11:27:25.794232 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 14 11:27:25 crc kubenswrapper[4904]: I0214 11:27:25.794677 4904 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 14 11:27:25 crc kubenswrapper[4904]: I0214 11:27:25.996009 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 14 11:27:28 crc kubenswrapper[4904]: I0214 11:27:28.267327 4904 generic.go:334] "Generic (PLEG): container finished" podID="c698a36a-d049-4fdb-b279-1b62d1f807d3" containerID="4d14917974e44826233222d2c4beb085c33e448861d26422e9fcec0bd9877e9a" exitCode=0 Feb 14 11:27:28 crc kubenswrapper[4904]: I0214 11:27:28.267415 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nnlq9" event={"ID":"c698a36a-d049-4fdb-b279-1b62d1f807d3","Type":"ContainerDied","Data":"4d14917974e44826233222d2c4beb085c33e448861d26422e9fcec0bd9877e9a"} Feb 14 11:27:28 crc kubenswrapper[4904]: I0214 11:27:28.780740 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6bc9d549b8-hdn6n" podUID="cfb252c0-ffb2-4d87-9c4e-b56cc2070d71" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 14 11:27:28 crc kubenswrapper[4904]: I0214 11:27:28.950601 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-687657b496-tz596" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.229981 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sflms" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.238618 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.290344 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nnlq9" event={"ID":"c698a36a-d049-4fdb-b279-1b62d1f807d3","Type":"ContainerDied","Data":"ed32c9db7ff9f74d2e4fdc9f70965da4f5969211cd1e7f3d9a5e4641d4515841"} Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.290384 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed32c9db7ff9f74d2e4fdc9f70965da4f5969211cd1e7f3d9a5e4641d4515841" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.290438 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nnlq9" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.291937 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sflms" event={"ID":"6b349fb1-5e71-45ff-88e9-03592bfe2504","Type":"ContainerDied","Data":"903e1842b32d1188981baed3b1e4cd5f7449328af56439f5dc20700bd5617a34"} Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.291960 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="903e1842b32d1188981baed3b1e4cd5f7449328af56439f5dc20700bd5617a34" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.292016 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sflms" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.372956 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b349fb1-5e71-45ff-88e9-03592bfe2504-combined-ca-bundle\") pod \"6b349fb1-5e71-45ff-88e9-03592bfe2504\" (UID: \"6b349fb1-5e71-45ff-88e9-03592bfe2504\") " Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.373048 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-db-sync-config-data\") pod \"c698a36a-d049-4fdb-b279-1b62d1f807d3\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.373092 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-scripts\") pod \"c698a36a-d049-4fdb-b279-1b62d1f807d3\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.373163 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf8c7\" (UniqueName: \"kubernetes.io/projected/c698a36a-d049-4fdb-b279-1b62d1f807d3-kube-api-access-bf8c7\") pod \"c698a36a-d049-4fdb-b279-1b62d1f807d3\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.373182 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4djz\" (UniqueName: \"kubernetes.io/projected/6b349fb1-5e71-45ff-88e9-03592bfe2504-kube-api-access-f4djz\") pod \"6b349fb1-5e71-45ff-88e9-03592bfe2504\" (UID: \"6b349fb1-5e71-45ff-88e9-03592bfe2504\") " Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.373211 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b349fb1-5e71-45ff-88e9-03592bfe2504-db-sync-config-data\") pod \"6b349fb1-5e71-45ff-88e9-03592bfe2504\" (UID: \"6b349fb1-5e71-45ff-88e9-03592bfe2504\") " Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.373231 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-config-data\") pod \"c698a36a-d049-4fdb-b279-1b62d1f807d3\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.373272 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-combined-ca-bundle\") pod \"c698a36a-d049-4fdb-b279-1b62d1f807d3\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.373310 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c698a36a-d049-4fdb-b279-1b62d1f807d3-etc-machine-id\") pod \"c698a36a-d049-4fdb-b279-1b62d1f807d3\" (UID: \"c698a36a-d049-4fdb-b279-1b62d1f807d3\") " Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.373816 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c698a36a-d049-4fdb-b279-1b62d1f807d3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c698a36a-d049-4fdb-b279-1b62d1f807d3" (UID: "c698a36a-d049-4fdb-b279-1b62d1f807d3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.379412 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-scripts" (OuterVolumeSpecName: "scripts") pod "c698a36a-d049-4fdb-b279-1b62d1f807d3" (UID: "c698a36a-d049-4fdb-b279-1b62d1f807d3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.379641 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b349fb1-5e71-45ff-88e9-03592bfe2504-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6b349fb1-5e71-45ff-88e9-03592bfe2504" (UID: "6b349fb1-5e71-45ff-88e9-03592bfe2504"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.379912 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c698a36a-d049-4fdb-b279-1b62d1f807d3-kube-api-access-bf8c7" (OuterVolumeSpecName: "kube-api-access-bf8c7") pod "c698a36a-d049-4fdb-b279-1b62d1f807d3" (UID: "c698a36a-d049-4fdb-b279-1b62d1f807d3"). InnerVolumeSpecName "kube-api-access-bf8c7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.380994 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b349fb1-5e71-45ff-88e9-03592bfe2504-kube-api-access-f4djz" (OuterVolumeSpecName: "kube-api-access-f4djz") pod "6b349fb1-5e71-45ff-88e9-03592bfe2504" (UID: "6b349fb1-5e71-45ff-88e9-03592bfe2504"). InnerVolumeSpecName "kube-api-access-f4djz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.383624 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c698a36a-d049-4fdb-b279-1b62d1f807d3" (UID: "c698a36a-d049-4fdb-b279-1b62d1f807d3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.421517 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c698a36a-d049-4fdb-b279-1b62d1f807d3" (UID: "c698a36a-d049-4fdb-b279-1b62d1f807d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.452096 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-config-data" (OuterVolumeSpecName: "config-data") pod "c698a36a-d049-4fdb-b279-1b62d1f807d3" (UID: "c698a36a-d049-4fdb-b279-1b62d1f807d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.468093 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b349fb1-5e71-45ff-88e9-03592bfe2504-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b349fb1-5e71-45ff-88e9-03592bfe2504" (UID: "6b349fb1-5e71-45ff-88e9-03592bfe2504"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.481741 4904 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b349fb1-5e71-45ff-88e9-03592bfe2504-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.481789 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.481803 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.481815 4904 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c698a36a-d049-4fdb-b279-1b62d1f807d3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.481827 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b349fb1-5e71-45ff-88e9-03592bfe2504-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.481862 4904 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.481872 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c698a36a-d049-4fdb-b279-1b62d1f807d3-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.481884 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf8c7\" (UniqueName: \"kubernetes.io/projected/c698a36a-d049-4fdb-b279-1b62d1f807d3-kube-api-access-bf8c7\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.481898 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4djz\" (UniqueName: \"kubernetes.io/projected/6b349fb1-5e71-45ff-88e9-03592bfe2504-kube-api-access-f4djz\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.601399 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 14 11:27:30 crc kubenswrapper[4904]: E0214 11:27:30.601750 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b349fb1-5e71-45ff-88e9-03592bfe2504" containerName="barbican-db-sync" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.601767 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b349fb1-5e71-45ff-88e9-03592bfe2504" containerName="barbican-db-sync" Feb 14 11:27:30 crc kubenswrapper[4904]: E0214 11:27:30.601797 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c698a36a-d049-4fdb-b279-1b62d1f807d3" containerName="cinder-db-sync" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.601803 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="c698a36a-d049-4fdb-b279-1b62d1f807d3" containerName="cinder-db-sync" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.601983 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b349fb1-5e71-45ff-88e9-03592bfe2504" containerName="barbican-db-sync" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.602004 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="c698a36a-d049-4fdb-b279-1b62d1f807d3" containerName="cinder-db-sync" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.602907 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.612329 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-rvw8l" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.612569 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.612739 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.612856 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.673975 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.701166 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twqp4\" (UniqueName: \"kubernetes.io/projected/56cfc647-7ed2-4649-979b-8a3cd60bd44f-kube-api-access-twqp4\") pod \"cinder-scheduler-0\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.701218 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.701264 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.701312 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-scripts\") pod \"cinder-scheduler-0\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.701333 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-config-data\") pod \"cinder-scheduler-0\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.701380 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/56cfc647-7ed2-4649-979b-8a3cd60bd44f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.769261 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b895b5785-c99jr"] Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.772975 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.802203 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twqp4\" (UniqueName: \"kubernetes.io/projected/56cfc647-7ed2-4649-979b-8a3cd60bd44f-kube-api-access-twqp4\") pod \"cinder-scheduler-0\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.802257 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.802305 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.802343 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-scripts\") pod \"cinder-scheduler-0\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.802359 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-config-data\") pod \"cinder-scheduler-0\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.802398 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/56cfc647-7ed2-4649-979b-8a3cd60bd44f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.802471 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/56cfc647-7ed2-4649-979b-8a3cd60bd44f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.809253 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.809413 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-c99jr"] Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.810508 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-scripts\") pod \"cinder-scheduler-0\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.812009 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-config-data\") pod \"cinder-scheduler-0\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.816962 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.840095 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twqp4\" (UniqueName: \"kubernetes.io/projected/56cfc647-7ed2-4649-979b-8a3cd60bd44f-kube-api-access-twqp4\") pod \"cinder-scheduler-0\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.906022 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8z9m\" (UniqueName: \"kubernetes.io/projected/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-kube-api-access-g8z9m\") pod \"dnsmasq-dns-b895b5785-c99jr\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.906366 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-c99jr\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.906412 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-c99jr\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.906429 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-c99jr\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.906454 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-dns-svc\") pod \"dnsmasq-dns-b895b5785-c99jr\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.906585 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-config\") pod \"dnsmasq-dns-b895b5785-c99jr\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.931526 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.933363 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.935308 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.937968 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 14 11:27:30 crc kubenswrapper[4904]: I0214 11:27:30.944781 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.008471 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8z9m\" (UniqueName: \"kubernetes.io/projected/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-kube-api-access-g8z9m\") pod \"dnsmasq-dns-b895b5785-c99jr\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.008744 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-c99jr\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.008951 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-c99jr\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.009065 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-c99jr\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.009168 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-dns-svc\") pod \"dnsmasq-dns-b895b5785-c99jr\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.009269 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-config\") pod \"dnsmasq-dns-b895b5785-c99jr\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.010457 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-config\") pod \"dnsmasq-dns-b895b5785-c99jr\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.011491 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-c99jr\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.012236 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-c99jr\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.013049 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-c99jr\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.013137 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-dns-svc\") pod \"dnsmasq-dns-b895b5785-c99jr\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.035032 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8z9m\" (UniqueName: \"kubernetes.io/projected/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-kube-api-access-g8z9m\") pod \"dnsmasq-dns-b895b5785-c99jr\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.101410 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.111370 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.111427 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drlvn\" (UniqueName: \"kubernetes.io/projected/77818f86-c391-47b6-afff-5bf2831b9328-kube-api-access-drlvn\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.111460 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-config-data-custom\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.111488 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-scripts\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.111545 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-config-data\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.111593 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77818f86-c391-47b6-afff-5bf2831b9328-logs\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.111636 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/77818f86-c391-47b6-afff-5bf2831b9328-etc-machine-id\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.213700 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-config-data\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.213774 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77818f86-c391-47b6-afff-5bf2831b9328-logs\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.213812 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/77818f86-c391-47b6-afff-5bf2831b9328-etc-machine-id\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.213878 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.213907 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drlvn\" (UniqueName: \"kubernetes.io/projected/77818f86-c391-47b6-afff-5bf2831b9328-kube-api-access-drlvn\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.213930 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-config-data-custom\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.213947 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-scripts\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.215057 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77818f86-c391-47b6-afff-5bf2831b9328-logs\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.215136 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/77818f86-c391-47b6-afff-5bf2831b9328-etc-machine-id\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.222201 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-scripts\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.223255 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-config-data\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.228538 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-config-data-custom\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.233528 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.234443 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drlvn\" (UniqueName: \"kubernetes.io/projected/77818f86-c391-47b6-afff-5bf2831b9328-kube-api-access-drlvn\") pod \"cinder-api-0\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.252252 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.460378 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-686dcf6685-s975t"] Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.461713 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-686dcf6685-s975t" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.472206 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.480364 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-glj86" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.482843 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.490772 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-686dcf6685-s975t"] Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.511095 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6dfd997684-xhxq7"] Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.521604 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.525091 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.585735 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6dfd997684-xhxq7"] Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.625729 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f74e923-d64a-4161-9839-96a80dd93c2b-combined-ca-bundle\") pod \"barbican-keystone-listener-6dfd997684-xhxq7\" (UID: \"2f74e923-d64a-4161-9839-96a80dd93c2b\") " pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.625813 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d096f44e-6631-414c-a590-e9e5f2f0cde2-logs\") pod \"barbican-worker-686dcf6685-s975t\" (UID: \"d096f44e-6631-414c-a590-e9e5f2f0cde2\") " pod="openstack/barbican-worker-686dcf6685-s975t" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.637090 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsklt\" (UniqueName: \"kubernetes.io/projected/d096f44e-6631-414c-a590-e9e5f2f0cde2-kube-api-access-wsklt\") pod \"barbican-worker-686dcf6685-s975t\" (UID: \"d096f44e-6631-414c-a590-e9e5f2f0cde2\") " pod="openstack/barbican-worker-686dcf6685-s975t" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.637165 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqdgh\" (UniqueName: \"kubernetes.io/projected/2f74e923-d64a-4161-9839-96a80dd93c2b-kube-api-access-tqdgh\") pod \"barbican-keystone-listener-6dfd997684-xhxq7\" (UID: \"2f74e923-d64a-4161-9839-96a80dd93c2b\") " pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.637322 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f74e923-d64a-4161-9839-96a80dd93c2b-config-data-custom\") pod \"barbican-keystone-listener-6dfd997684-xhxq7\" (UID: \"2f74e923-d64a-4161-9839-96a80dd93c2b\") " pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.637361 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d096f44e-6631-414c-a590-e9e5f2f0cde2-config-data-custom\") pod \"barbican-worker-686dcf6685-s975t\" (UID: \"d096f44e-6631-414c-a590-e9e5f2f0cde2\") " pod="openstack/barbican-worker-686dcf6685-s975t" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.637394 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d096f44e-6631-414c-a590-e9e5f2f0cde2-config-data\") pod \"barbican-worker-686dcf6685-s975t\" (UID: \"d096f44e-6631-414c-a590-e9e5f2f0cde2\") " pod="openstack/barbican-worker-686dcf6685-s975t" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.637433 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d096f44e-6631-414c-a590-e9e5f2f0cde2-combined-ca-bundle\") pod \"barbican-worker-686dcf6685-s975t\" (UID: \"d096f44e-6631-414c-a590-e9e5f2f0cde2\") " pod="openstack/barbican-worker-686dcf6685-s975t" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.637497 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f74e923-d64a-4161-9839-96a80dd93c2b-logs\") pod \"barbican-keystone-listener-6dfd997684-xhxq7\" (UID: \"2f74e923-d64a-4161-9839-96a80dd93c2b\") " pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.637531 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f74e923-d64a-4161-9839-96a80dd93c2b-config-data\") pod \"barbican-keystone-listener-6dfd997684-xhxq7\" (UID: \"2f74e923-d64a-4161-9839-96a80dd93c2b\") " pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.744646 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d096f44e-6631-414c-a590-e9e5f2f0cde2-logs\") pod \"barbican-worker-686dcf6685-s975t\" (UID: \"d096f44e-6631-414c-a590-e9e5f2f0cde2\") " pod="openstack/barbican-worker-686dcf6685-s975t" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.744794 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsklt\" (UniqueName: \"kubernetes.io/projected/d096f44e-6631-414c-a590-e9e5f2f0cde2-kube-api-access-wsklt\") pod \"barbican-worker-686dcf6685-s975t\" (UID: \"d096f44e-6631-414c-a590-e9e5f2f0cde2\") " pod="openstack/barbican-worker-686dcf6685-s975t" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.744853 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqdgh\" (UniqueName: \"kubernetes.io/projected/2f74e923-d64a-4161-9839-96a80dd93c2b-kube-api-access-tqdgh\") pod \"barbican-keystone-listener-6dfd997684-xhxq7\" (UID: \"2f74e923-d64a-4161-9839-96a80dd93c2b\") " pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.744988 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f74e923-d64a-4161-9839-96a80dd93c2b-config-data-custom\") pod \"barbican-keystone-listener-6dfd997684-xhxq7\" (UID: \"2f74e923-d64a-4161-9839-96a80dd93c2b\") " pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.745025 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d096f44e-6631-414c-a590-e9e5f2f0cde2-config-data-custom\") pod \"barbican-worker-686dcf6685-s975t\" (UID: \"d096f44e-6631-414c-a590-e9e5f2f0cde2\") " pod="openstack/barbican-worker-686dcf6685-s975t" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.745062 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d096f44e-6631-414c-a590-e9e5f2f0cde2-config-data\") pod \"barbican-worker-686dcf6685-s975t\" (UID: \"d096f44e-6631-414c-a590-e9e5f2f0cde2\") " pod="openstack/barbican-worker-686dcf6685-s975t" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.745092 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d096f44e-6631-414c-a590-e9e5f2f0cde2-combined-ca-bundle\") pod \"barbican-worker-686dcf6685-s975t\" (UID: \"d096f44e-6631-414c-a590-e9e5f2f0cde2\") " pod="openstack/barbican-worker-686dcf6685-s975t" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.745150 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f74e923-d64a-4161-9839-96a80dd93c2b-logs\") pod \"barbican-keystone-listener-6dfd997684-xhxq7\" (UID: \"2f74e923-d64a-4161-9839-96a80dd93c2b\") " pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.745186 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f74e923-d64a-4161-9839-96a80dd93c2b-config-data\") pod \"barbican-keystone-listener-6dfd997684-xhxq7\" (UID: \"2f74e923-d64a-4161-9839-96a80dd93c2b\") " pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.745244 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f74e923-d64a-4161-9839-96a80dd93c2b-combined-ca-bundle\") pod \"barbican-keystone-listener-6dfd997684-xhxq7\" (UID: \"2f74e923-d64a-4161-9839-96a80dd93c2b\") " pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.746745 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d096f44e-6631-414c-a590-e9e5f2f0cde2-logs\") pod \"barbican-worker-686dcf6685-s975t\" (UID: \"d096f44e-6631-414c-a590-e9e5f2f0cde2\") " pod="openstack/barbican-worker-686dcf6685-s975t" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.789935 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f74e923-d64a-4161-9839-96a80dd93c2b-logs\") pod \"barbican-keystone-listener-6dfd997684-xhxq7\" (UID: \"2f74e923-d64a-4161-9839-96a80dd93c2b\") " pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.790705 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f74e923-d64a-4161-9839-96a80dd93c2b-config-data-custom\") pod \"barbican-keystone-listener-6dfd997684-xhxq7\" (UID: \"2f74e923-d64a-4161-9839-96a80dd93c2b\") " pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.795243 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d096f44e-6631-414c-a590-e9e5f2f0cde2-config-data-custom\") pod \"barbican-worker-686dcf6685-s975t\" (UID: \"d096f44e-6631-414c-a590-e9e5f2f0cde2\") " pod="openstack/barbican-worker-686dcf6685-s975t" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.806937 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-c99jr"] Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.811112 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d096f44e-6631-414c-a590-e9e5f2f0cde2-combined-ca-bundle\") pod \"barbican-worker-686dcf6685-s975t\" (UID: \"d096f44e-6631-414c-a590-e9e5f2f0cde2\") " pod="openstack/barbican-worker-686dcf6685-s975t" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.822045 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f74e923-d64a-4161-9839-96a80dd93c2b-config-data\") pod \"barbican-keystone-listener-6dfd997684-xhxq7\" (UID: \"2f74e923-d64a-4161-9839-96a80dd93c2b\") " pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.823390 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqdgh\" (UniqueName: \"kubernetes.io/projected/2f74e923-d64a-4161-9839-96a80dd93c2b-kube-api-access-tqdgh\") pod \"barbican-keystone-listener-6dfd997684-xhxq7\" (UID: \"2f74e923-d64a-4161-9839-96a80dd93c2b\") " pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.829231 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f74e923-d64a-4161-9839-96a80dd93c2b-combined-ca-bundle\") pod \"barbican-keystone-listener-6dfd997684-xhxq7\" (UID: \"2f74e923-d64a-4161-9839-96a80dd93c2b\") " pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.834600 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-s54q7"] Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.836504 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.842261 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsklt\" (UniqueName: \"kubernetes.io/projected/d096f44e-6631-414c-a590-e9e5f2f0cde2-kube-api-access-wsklt\") pod \"barbican-worker-686dcf6685-s975t\" (UID: \"d096f44e-6631-414c-a590-e9e5f2f0cde2\") " pod="openstack/barbican-worker-686dcf6685-s975t" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.858144 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d096f44e-6631-414c-a590-e9e5f2f0cde2-config-data\") pod \"barbican-worker-686dcf6685-s975t\" (UID: \"d096f44e-6631-414c-a590-e9e5f2f0cde2\") " pod="openstack/barbican-worker-686dcf6685-s975t" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.858574 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.871609 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-s54q7"] Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.882471 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-797ff5c494-4sfjf"] Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.904498 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.918212 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.930033 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-797ff5c494-4sfjf"] Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.948550 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-config\") pod \"dnsmasq-dns-5c9776ccc5-s54q7\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.948778 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghwtp\" (UniqueName: \"kubernetes.io/projected/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-kube-api-access-ghwtp\") pod \"dnsmasq-dns-5c9776ccc5-s54q7\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.948903 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-s54q7\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.949054 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-s54q7\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.949112 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-s54q7\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:31 crc kubenswrapper[4904]: I0214 11:27:31.949171 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-s54q7\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.050762 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b12a8202-af0b-403e-9767-7cbaeb661cd5-combined-ca-bundle\") pod \"barbican-api-797ff5c494-4sfjf\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.050814 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghwtp\" (UniqueName: \"kubernetes.io/projected/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-kube-api-access-ghwtp\") pod \"dnsmasq-dns-5c9776ccc5-s54q7\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.050847 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b12a8202-af0b-403e-9767-7cbaeb661cd5-config-data\") pod \"barbican-api-797ff5c494-4sfjf\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.050898 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-s54q7\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.050920 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b12a8202-af0b-403e-9767-7cbaeb661cd5-config-data-custom\") pod \"barbican-api-797ff5c494-4sfjf\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.050972 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-s54q7\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.050999 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-s54q7\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.051023 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-s54q7\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.051054 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-config\") pod \"dnsmasq-dns-5c9776ccc5-s54q7\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.051076 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b12a8202-af0b-403e-9767-7cbaeb661cd5-logs\") pod \"barbican-api-797ff5c494-4sfjf\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.051115 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-974wd\" (UniqueName: \"kubernetes.io/projected/b12a8202-af0b-403e-9767-7cbaeb661cd5-kube-api-access-974wd\") pod \"barbican-api-797ff5c494-4sfjf\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.051978 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-s54q7\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.052100 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-s54q7\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.052350 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-s54q7\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.052681 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-s54q7\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.053061 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-config\") pod \"dnsmasq-dns-5c9776ccc5-s54q7\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.069883 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghwtp\" (UniqueName: \"kubernetes.io/projected/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-kube-api-access-ghwtp\") pod \"dnsmasq-dns-5c9776ccc5-s54q7\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.126813 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-686dcf6685-s975t" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.152724 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-974wd\" (UniqueName: \"kubernetes.io/projected/b12a8202-af0b-403e-9767-7cbaeb661cd5-kube-api-access-974wd\") pod \"barbican-api-797ff5c494-4sfjf\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.152785 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b12a8202-af0b-403e-9767-7cbaeb661cd5-combined-ca-bundle\") pod \"barbican-api-797ff5c494-4sfjf\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.152821 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b12a8202-af0b-403e-9767-7cbaeb661cd5-config-data\") pod \"barbican-api-797ff5c494-4sfjf\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.152878 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b12a8202-af0b-403e-9767-7cbaeb661cd5-config-data-custom\") pod \"barbican-api-797ff5c494-4sfjf\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.152979 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b12a8202-af0b-403e-9767-7cbaeb661cd5-logs\") pod \"barbican-api-797ff5c494-4sfjf\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.153414 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b12a8202-af0b-403e-9767-7cbaeb661cd5-logs\") pod \"barbican-api-797ff5c494-4sfjf\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.157462 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b12a8202-af0b-403e-9767-7cbaeb661cd5-combined-ca-bundle\") pod \"barbican-api-797ff5c494-4sfjf\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.157624 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b12a8202-af0b-403e-9767-7cbaeb661cd5-config-data-custom\") pod \"barbican-api-797ff5c494-4sfjf\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.162173 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b12a8202-af0b-403e-9767-7cbaeb661cd5-config-data\") pod \"barbican-api-797ff5c494-4sfjf\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.184090 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-974wd\" (UniqueName: \"kubernetes.io/projected/b12a8202-af0b-403e-9767-7cbaeb661cd5-kube-api-access-974wd\") pod \"barbican-api-797ff5c494-4sfjf\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.239192 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:32 crc kubenswrapper[4904]: I0214 11:27:32.249024 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:33 crc kubenswrapper[4904]: E0214 11:27:33.326251 4904 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Feb 14 11:27:33 crc kubenswrapper[4904]: E0214 11:27:33.326552 4904 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xgmqf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(16339330-a7b5-471a-ad7e-1a8d4627c3e2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 14 11:27:33 crc kubenswrapper[4904]: E0214 11:27:33.327949 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="16339330-a7b5-471a-ad7e-1a8d4627c3e2" Feb 14 11:27:33 crc kubenswrapper[4904]: I0214 11:27:33.665941 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 14 11:27:33 crc kubenswrapper[4904]: I0214 11:27:33.967595 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 14 11:27:34 crc kubenswrapper[4904]: I0214 11:27:34.146383 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 14 11:27:34 crc kubenswrapper[4904]: W0214 11:27:34.166074 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77818f86_c391_47b6_afff_5bf2831b9328.slice/crio-f853eafe1f5b6bee6df9fa4477ebdb1430c7b9d885ce5bcdbcea02ebe3f335d9 WatchSource:0}: Error finding container f853eafe1f5b6bee6df9fa4477ebdb1430c7b9d885ce5bcdbcea02ebe3f335d9: Status 404 returned error can't find the container with id f853eafe1f5b6bee6df9fa4477ebdb1430c7b9d885ce5bcdbcea02ebe3f335d9 Feb 14 11:27:34 crc kubenswrapper[4904]: I0214 11:27:34.210230 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-c99jr"] Feb 14 11:27:34 crc kubenswrapper[4904]: I0214 11:27:34.321788 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-686dcf6685-s975t"] Feb 14 11:27:34 crc kubenswrapper[4904]: I0214 11:27:34.351079 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-s54q7"] Feb 14 11:27:34 crc kubenswrapper[4904]: I0214 11:27:34.375251 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"77818f86-c391-47b6-afff-5bf2831b9328","Type":"ContainerStarted","Data":"f853eafe1f5b6bee6df9fa4477ebdb1430c7b9d885ce5bcdbcea02ebe3f335d9"} Feb 14 11:27:34 crc kubenswrapper[4904]: I0214 11:27:34.376776 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b895b5785-c99jr" event={"ID":"e92a9c71-c9f6-45d4-a6ec-f031a45a1675","Type":"ContainerStarted","Data":"02039d4359af1ddf8609682a1bc4fc3f05d6b7a7c1b98f85a673eec9a179f6df"} Feb 14 11:27:34 crc kubenswrapper[4904]: I0214 11:27:34.381580 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="16339330-a7b5-471a-ad7e-1a8d4627c3e2" containerName="ceilometer-notification-agent" containerID="cri-o://c9dc77b78a504691a7a3af90a1868fc0b6ba517a927791a09b98aab71a44ff6d" gracePeriod=30 Feb 14 11:27:34 crc kubenswrapper[4904]: I0214 11:27:34.381824 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="16339330-a7b5-471a-ad7e-1a8d4627c3e2" containerName="sg-core" containerID="cri-o://7698083f123e53016f0ffaa8355298315cb001448a6b1604956e2988bba73597" gracePeriod=30 Feb 14 11:27:34 crc kubenswrapper[4904]: I0214 11:27:34.384913 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"56cfc647-7ed2-4649-979b-8a3cd60bd44f","Type":"ContainerStarted","Data":"d6370b6c19b5bc0cae9cf3324da385a7898ed64417ddbe121d1235f21bfd6886"} Feb 14 11:27:34 crc kubenswrapper[4904]: I0214 11:27:34.390341 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6dfd997684-xhxq7"] Feb 14 11:27:34 crc kubenswrapper[4904]: I0214 11:27:34.536145 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-797ff5c494-4sfjf"] Feb 14 11:27:35 crc kubenswrapper[4904]: I0214 11:27:35.415401 4904 generic.go:334] "Generic (PLEG): container finished" podID="0d09a98a-7e3d-405a-b2d1-6f6537f4c940" containerID="73797f3df06ae349b13be6fb4aa32a1c40834ec93034f0999fb191bcfaaa20bd" exitCode=0 Feb 14 11:27:35 crc kubenswrapper[4904]: I0214 11:27:35.415922 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" event={"ID":"0d09a98a-7e3d-405a-b2d1-6f6537f4c940","Type":"ContainerDied","Data":"73797f3df06ae349b13be6fb4aa32a1c40834ec93034f0999fb191bcfaaa20bd"} Feb 14 11:27:35 crc kubenswrapper[4904]: I0214 11:27:35.415954 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" event={"ID":"0d09a98a-7e3d-405a-b2d1-6f6537f4c940","Type":"ContainerStarted","Data":"84522020ca1948cf4c8dbca5113534cc75e06bf03ae4a2e083e3569883c167e7"} Feb 14 11:27:35 crc kubenswrapper[4904]: I0214 11:27:35.447211 4904 generic.go:334] "Generic (PLEG): container finished" podID="16339330-a7b5-471a-ad7e-1a8d4627c3e2" containerID="7698083f123e53016f0ffaa8355298315cb001448a6b1604956e2988bba73597" exitCode=2 Feb 14 11:27:35 crc kubenswrapper[4904]: I0214 11:27:35.447238 4904 generic.go:334] "Generic (PLEG): container finished" podID="16339330-a7b5-471a-ad7e-1a8d4627c3e2" containerID="c9dc77b78a504691a7a3af90a1868fc0b6ba517a927791a09b98aab71a44ff6d" exitCode=0 Feb 14 11:27:35 crc kubenswrapper[4904]: I0214 11:27:35.447306 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16339330-a7b5-471a-ad7e-1a8d4627c3e2","Type":"ContainerDied","Data":"7698083f123e53016f0ffaa8355298315cb001448a6b1604956e2988bba73597"} Feb 14 11:27:35 crc kubenswrapper[4904]: I0214 11:27:35.447331 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16339330-a7b5-471a-ad7e-1a8d4627c3e2","Type":"ContainerDied","Data":"c9dc77b78a504691a7a3af90a1868fc0b6ba517a927791a09b98aab71a44ff6d"} Feb 14 11:27:35 crc kubenswrapper[4904]: I0214 11:27:35.448513 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-686dcf6685-s975t" event={"ID":"d096f44e-6631-414c-a590-e9e5f2f0cde2","Type":"ContainerStarted","Data":"c52c31dcdc9e3e968e0791bd6667bad8dffbbd170db4a7e9589cbd3797ef0666"} Feb 14 11:27:35 crc kubenswrapper[4904]: I0214 11:27:35.486265 4904 generic.go:334] "Generic (PLEG): container finished" podID="e92a9c71-c9f6-45d4-a6ec-f031a45a1675" containerID="9da8a6b74a1c2378a8dc8c341fa8c9340c79ff3b2d87a752f406c6a5f416ad06" exitCode=0 Feb 14 11:27:35 crc kubenswrapper[4904]: I0214 11:27:35.486364 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b895b5785-c99jr" event={"ID":"e92a9c71-c9f6-45d4-a6ec-f031a45a1675","Type":"ContainerDied","Data":"9da8a6b74a1c2378a8dc8c341fa8c9340c79ff3b2d87a752f406c6a5f416ad06"} Feb 14 11:27:35 crc kubenswrapper[4904]: I0214 11:27:35.573073 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-797ff5c494-4sfjf" event={"ID":"b12a8202-af0b-403e-9767-7cbaeb661cd5","Type":"ContainerStarted","Data":"c8a3039de6035c530e1e87fcfbe0392cf888fdba457441ecf5e0475fb63ac5c7"} Feb 14 11:27:35 crc kubenswrapper[4904]: I0214 11:27:35.573113 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-797ff5c494-4sfjf" event={"ID":"b12a8202-af0b-403e-9767-7cbaeb661cd5","Type":"ContainerStarted","Data":"fd08f31feb180fc1a7ce12ed8a4dae1b50ad042c55310211dfd66ab94d9a0eee"} Feb 14 11:27:35 crc kubenswrapper[4904]: I0214 11:27:35.586978 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" event={"ID":"2f74e923-d64a-4161-9839-96a80dd93c2b","Type":"ContainerStarted","Data":"898c1ba710c228c4cc119c08671966fe07f7703cb52578e8969b86271a7d3557"} Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.300235 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.384477 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.492280 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16339330-a7b5-471a-ad7e-1a8d4627c3e2-log-httpd\") pod \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.492350 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-combined-ca-bundle\") pod \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.492446 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-ovsdbserver-nb\") pod \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.492497 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8z9m\" (UniqueName: \"kubernetes.io/projected/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-kube-api-access-g8z9m\") pod \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.492527 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-sg-core-conf-yaml\") pod \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.492545 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-ovsdbserver-sb\") pod \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.492659 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-dns-swift-storage-0\") pod \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.492650 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16339330-a7b5-471a-ad7e-1a8d4627c3e2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "16339330-a7b5-471a-ad7e-1a8d4627c3e2" (UID: "16339330-a7b5-471a-ad7e-1a8d4627c3e2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.492680 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-config-data\") pod \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.492743 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-scripts\") pod \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.492769 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgmqf\" (UniqueName: \"kubernetes.io/projected/16339330-a7b5-471a-ad7e-1a8d4627c3e2-kube-api-access-xgmqf\") pod \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.492810 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-config\") pod \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.492853 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-dns-svc\") pod \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\" (UID: \"e92a9c71-c9f6-45d4-a6ec-f031a45a1675\") " Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.492936 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16339330-a7b5-471a-ad7e-1a8d4627c3e2-run-httpd\") pod \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\" (UID: \"16339330-a7b5-471a-ad7e-1a8d4627c3e2\") " Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.493416 4904 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16339330-a7b5-471a-ad7e-1a8d4627c3e2-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.493764 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16339330-a7b5-471a-ad7e-1a8d4627c3e2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "16339330-a7b5-471a-ad7e-1a8d4627c3e2" (UID: "16339330-a7b5-471a-ad7e-1a8d4627c3e2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.515992 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-scripts" (OuterVolumeSpecName: "scripts") pod "16339330-a7b5-471a-ad7e-1a8d4627c3e2" (UID: "16339330-a7b5-471a-ad7e-1a8d4627c3e2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.516875 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-kube-api-access-g8z9m" (OuterVolumeSpecName: "kube-api-access-g8z9m") pod "e92a9c71-c9f6-45d4-a6ec-f031a45a1675" (UID: "e92a9c71-c9f6-45d4-a6ec-f031a45a1675"). InnerVolumeSpecName "kube-api-access-g8z9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.518968 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16339330-a7b5-471a-ad7e-1a8d4627c3e2-kube-api-access-xgmqf" (OuterVolumeSpecName: "kube-api-access-xgmqf") pod "16339330-a7b5-471a-ad7e-1a8d4627c3e2" (UID: "16339330-a7b5-471a-ad7e-1a8d4627c3e2"). InnerVolumeSpecName "kube-api-access-xgmqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.552808 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e92a9c71-c9f6-45d4-a6ec-f031a45a1675" (UID: "e92a9c71-c9f6-45d4-a6ec-f031a45a1675"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.566091 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.583773 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e92a9c71-c9f6-45d4-a6ec-f031a45a1675" (UID: "e92a9c71-c9f6-45d4-a6ec-f031a45a1675"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.599190 4904 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16339330-a7b5-471a-ad7e-1a8d4627c3e2-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.599228 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.599241 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8z9m\" (UniqueName: \"kubernetes.io/projected/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-kube-api-access-g8z9m\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.599252 4904 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.599262 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.599273 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgmqf\" (UniqueName: \"kubernetes.io/projected/16339330-a7b5-471a-ad7e-1a8d4627c3e2-kube-api-access-xgmqf\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.641460 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16339330-a7b5-471a-ad7e-1a8d4627c3e2" (UID: "16339330-a7b5-471a-ad7e-1a8d4627c3e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.643358 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-config" (OuterVolumeSpecName: "config") pod "e92a9c71-c9f6-45d4-a6ec-f031a45a1675" (UID: "e92a9c71-c9f6-45d4-a6ec-f031a45a1675"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.659581 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-config-data" (OuterVolumeSpecName: "config-data") pod "16339330-a7b5-471a-ad7e-1a8d4627c3e2" (UID: "16339330-a7b5-471a-ad7e-1a8d4627c3e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.681717 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e92a9c71-c9f6-45d4-a6ec-f031a45a1675" (UID: "e92a9c71-c9f6-45d4-a6ec-f031a45a1675"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.691009 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "16339330-a7b5-471a-ad7e-1a8d4627c3e2" (UID: "16339330-a7b5-471a-ad7e-1a8d4627c3e2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.702070 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"56cfc647-7ed2-4649-979b-8a3cd60bd44f","Type":"ContainerStarted","Data":"22715d0f5fd42bf2706df319554222c8e0581c422c1f7dcaa7186748f5c37c3e"} Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.703395 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.703416 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.703426 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.703436 4904 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16339330-a7b5-471a-ad7e-1a8d4627c3e2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.703445 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.712202 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e92a9c71-c9f6-45d4-a6ec-f031a45a1675" (UID: "e92a9c71-c9f6-45d4-a6ec-f031a45a1675"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.730117 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"77818f86-c391-47b6-afff-5bf2831b9328","Type":"ContainerStarted","Data":"e07211442907acafa7321466c70ea387e33038cb4a3b210324aa5761de4e6ff1"} Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.765726 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" event={"ID":"0d09a98a-7e3d-405a-b2d1-6f6537f4c940","Type":"ContainerStarted","Data":"a0f74c1dcef80c1649018c3d8c6f171033515a1a167abcbd736a974e775268cc"} Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.765891 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.809712 4904 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e92a9c71-c9f6-45d4-a6ec-f031a45a1675-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.823830 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16339330-a7b5-471a-ad7e-1a8d4627c3e2","Type":"ContainerDied","Data":"deb50cca9f879e4b9ec57f6beab522df24ed4ba1a8d60a33e487fe0676797254"} Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.823892 4904 scope.go:117] "RemoveContainer" containerID="7698083f123e53016f0ffaa8355298315cb001448a6b1604956e2988bba73597" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.824010 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.876448 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-c99jr" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.876719 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b895b5785-c99jr" event={"ID":"e92a9c71-c9f6-45d4-a6ec-f031a45a1675","Type":"ContainerDied","Data":"02039d4359af1ddf8609682a1bc4fc3f05d6b7a7c1b98f85a673eec9a179f6df"} Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.953589 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-797ff5c494-4sfjf" event={"ID":"b12a8202-af0b-403e-9767-7cbaeb661cd5","Type":"ContainerStarted","Data":"9dfd9c5085ab7919e13381193642df7de78c9d7afa7829b6627ba596f57c0a72"} Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.958392 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.958443 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:36 crc kubenswrapper[4904]: I0214 11:27:36.985641 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" podStartSLOduration=5.985616484 podStartE2EDuration="5.985616484s" podCreationTimestamp="2026-02-14 11:27:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:27:36.810314723 +0000 UTC m=+1047.623079384" watchObservedRunningTime="2026-02-14 11:27:36.985616484 +0000 UTC m=+1047.798381145" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.108148 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.125584 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.141846 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-797ff5c494-4sfjf" podStartSLOduration=6.141813989 podStartE2EDuration="6.141813989s" podCreationTimestamp="2026-02-14 11:27:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:27:36.983304901 +0000 UTC m=+1047.796069562" watchObservedRunningTime="2026-02-14 11:27:37.141813989 +0000 UTC m=+1047.954578650" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.150983 4904 scope.go:117] "RemoveContainer" containerID="c9dc77b78a504691a7a3af90a1868fc0b6ba517a927791a09b98aab71a44ff6d" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.175346 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:27:37 crc kubenswrapper[4904]: E0214 11:27:37.175706 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16339330-a7b5-471a-ad7e-1a8d4627c3e2" containerName="ceilometer-notification-agent" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.175721 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="16339330-a7b5-471a-ad7e-1a8d4627c3e2" containerName="ceilometer-notification-agent" Feb 14 11:27:37 crc kubenswrapper[4904]: E0214 11:27:37.175737 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e92a9c71-c9f6-45d4-a6ec-f031a45a1675" containerName="init" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.175744 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="e92a9c71-c9f6-45d4-a6ec-f031a45a1675" containerName="init" Feb 14 11:27:37 crc kubenswrapper[4904]: E0214 11:27:37.175759 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16339330-a7b5-471a-ad7e-1a8d4627c3e2" containerName="sg-core" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.175765 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="16339330-a7b5-471a-ad7e-1a8d4627c3e2" containerName="sg-core" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.176046 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="16339330-a7b5-471a-ad7e-1a8d4627c3e2" containerName="sg-core" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.176067 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="e92a9c71-c9f6-45d4-a6ec-f031a45a1675" containerName="init" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.176082 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="16339330-a7b5-471a-ad7e-1a8d4627c3e2" containerName="ceilometer-notification-agent" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.177620 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.185083 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.188638 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.193828 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.225295 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-58c9948449-g7lbx"] Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.225859 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-58c9948449-g7lbx" podUID="4cc31d3d-1f06-48a1-9856-e1a997bfb11e" containerName="neutron-api" containerID="cri-o://df39d52890d4f62c453dfa0452f433a477f0e1ede96fc0bb86defbf9ea1c50a1" gracePeriod=30 Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.226534 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-58c9948449-g7lbx" podUID="4cc31d3d-1f06-48a1-9856-e1a997bfb11e" containerName="neutron-httpd" containerID="cri-o://eea579d5897474d21fdec00430fbfcb39d8f2d878d90ca0eb9305bd3657fc893" gracePeriod=30 Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.250655 4904 scope.go:117] "RemoveContainer" containerID="9da8a6b74a1c2378a8dc8c341fa8c9340c79ff3b2d87a752f406c6a5f416ad06" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.263271 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-c99jr"] Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.268404 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-c99jr"] Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.300983 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7856f6fd6f-4rhbv"] Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.304617 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.313292 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7856f6fd6f-4rhbv"] Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.334433 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bccb523-3f10-41ba-aeba-0cd9675d20e2-run-httpd\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.334479 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-scripts\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.334504 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27qbv\" (UniqueName: \"kubernetes.io/projected/3bccb523-3f10-41ba-aeba-0cd9675d20e2-kube-api-access-27qbv\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.334530 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.334571 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bccb523-3f10-41ba-aeba-0cd9675d20e2-log-httpd\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.334601 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.334644 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-config-data\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.349161 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-58c9948449-g7lbx" podUID="4cc31d3d-1f06-48a1-9856-e1a997bfb11e" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.158:9696/\": read tcp 10.217.0.2:44086->10.217.0.158:9696: read: connection reset by peer" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.436696 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8261bc0e-8af3-4000-aa2d-de7153a235ca-combined-ca-bundle\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.436747 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-config-data\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.436776 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8261bc0e-8af3-4000-aa2d-de7153a235ca-config\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.436799 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8261bc0e-8af3-4000-aa2d-de7153a235ca-internal-tls-certs\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.436814 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8261bc0e-8af3-4000-aa2d-de7153a235ca-httpd-config\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.436889 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x2tn\" (UniqueName: \"kubernetes.io/projected/8261bc0e-8af3-4000-aa2d-de7153a235ca-kube-api-access-8x2tn\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.436913 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-scripts\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.436927 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bccb523-3f10-41ba-aeba-0cd9675d20e2-run-httpd\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.436942 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8261bc0e-8af3-4000-aa2d-de7153a235ca-public-tls-certs\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.436959 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27qbv\" (UniqueName: \"kubernetes.io/projected/3bccb523-3f10-41ba-aeba-0cd9675d20e2-kube-api-access-27qbv\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.436977 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.437021 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bccb523-3f10-41ba-aeba-0cd9675d20e2-log-httpd\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.437038 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8261bc0e-8af3-4000-aa2d-de7153a235ca-ovndb-tls-certs\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.437057 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.445598 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bccb523-3f10-41ba-aeba-0cd9675d20e2-run-httpd\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.450000 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bccb523-3f10-41ba-aeba-0cd9675d20e2-log-httpd\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.460092 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.460680 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-scripts\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.461565 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-config-data\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.461912 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.464505 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27qbv\" (UniqueName: \"kubernetes.io/projected/3bccb523-3f10-41ba-aeba-0cd9675d20e2-kube-api-access-27qbv\") pod \"ceilometer-0\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.538737 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x2tn\" (UniqueName: \"kubernetes.io/projected/8261bc0e-8af3-4000-aa2d-de7153a235ca-kube-api-access-8x2tn\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.538793 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8261bc0e-8af3-4000-aa2d-de7153a235ca-public-tls-certs\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.538851 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8261bc0e-8af3-4000-aa2d-de7153a235ca-ovndb-tls-certs\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.538894 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8261bc0e-8af3-4000-aa2d-de7153a235ca-combined-ca-bundle\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.538927 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8261bc0e-8af3-4000-aa2d-de7153a235ca-config\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.538949 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8261bc0e-8af3-4000-aa2d-de7153a235ca-internal-tls-certs\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.538964 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8261bc0e-8af3-4000-aa2d-de7153a235ca-httpd-config\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.542622 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.547591 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8261bc0e-8af3-4000-aa2d-de7153a235ca-combined-ca-bundle\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.548228 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8261bc0e-8af3-4000-aa2d-de7153a235ca-public-tls-certs\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.549420 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8261bc0e-8af3-4000-aa2d-de7153a235ca-ovndb-tls-certs\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.554526 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8261bc0e-8af3-4000-aa2d-de7153a235ca-httpd-config\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.553543 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8261bc0e-8af3-4000-aa2d-de7153a235ca-config\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.564528 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8261bc0e-8af3-4000-aa2d-de7153a235ca-internal-tls-certs\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.570486 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x2tn\" (UniqueName: \"kubernetes.io/projected/8261bc0e-8af3-4000-aa2d-de7153a235ca-kube-api-access-8x2tn\") pod \"neutron-7856f6fd6f-4rhbv\" (UID: \"8261bc0e-8af3-4000-aa2d-de7153a235ca\") " pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.709473 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.847374 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16339330-a7b5-471a-ad7e-1a8d4627c3e2" path="/var/lib/kubelet/pods/16339330-a7b5-471a-ad7e-1a8d4627c3e2/volumes" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.848114 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e92a9c71-c9f6-45d4-a6ec-f031a45a1675" path="/var/lib/kubelet/pods/e92a9c71-c9f6-45d4-a6ec-f031a45a1675/volumes" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.978210 4904 generic.go:334] "Generic (PLEG): container finished" podID="4cc31d3d-1f06-48a1-9856-e1a997bfb11e" containerID="eea579d5897474d21fdec00430fbfcb39d8f2d878d90ca0eb9305bd3657fc893" exitCode=0 Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.978283 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58c9948449-g7lbx" event={"ID":"4cc31d3d-1f06-48a1-9856-e1a997bfb11e","Type":"ContainerDied","Data":"eea579d5897474d21fdec00430fbfcb39d8f2d878d90ca0eb9305bd3657fc893"} Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.985091 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"56cfc647-7ed2-4649-979b-8a3cd60bd44f","Type":"ContainerStarted","Data":"77445f189d44c809b28038a9734762f33fa07206f66b5250894d21888fd291ac"} Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.992575 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"77818f86-c391-47b6-afff-5bf2831b9328","Type":"ContainerStarted","Data":"dd3d3c3307a079686348e28c4f4826c7b4b8390ad9774ebf70d9ef1a082f06d6"} Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.993451 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="77818f86-c391-47b6-afff-5bf2831b9328" containerName="cinder-api-log" containerID="cri-o://e07211442907acafa7321466c70ea387e33038cb4a3b210324aa5761de4e6ff1" gracePeriod=30 Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.993630 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 14 11:27:37 crc kubenswrapper[4904]: I0214 11:27:37.993672 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="77818f86-c391-47b6-afff-5bf2831b9328" containerName="cinder-api" containerID="cri-o://dd3d3c3307a079686348e28c4f4826c7b4b8390ad9774ebf70d9ef1a082f06d6" gracePeriod=30 Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.025971 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=7.156922984 podStartE2EDuration="8.025953911s" podCreationTimestamp="2026-02-14 11:27:30 +0000 UTC" firstStartedPulling="2026-02-14 11:27:33.97435539 +0000 UTC m=+1044.787120051" lastFinishedPulling="2026-02-14 11:27:34.843386317 +0000 UTC m=+1045.656150978" observedRunningTime="2026-02-14 11:27:38.018231229 +0000 UTC m=+1048.830995890" watchObservedRunningTime="2026-02-14 11:27:38.025953911 +0000 UTC m=+1048.838718572" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.056954 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=8.056933903 podStartE2EDuration="8.056933903s" podCreationTimestamp="2026-02-14 11:27:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:27:38.045057777 +0000 UTC m=+1048.857822438" watchObservedRunningTime="2026-02-14 11:27:38.056933903 +0000 UTC m=+1048.869698564" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.077661 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-c9489b69d-msqgc"] Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.079162 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.082165 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.082957 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.101481 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c9489b69d-msqgc"] Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.152667 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd5faf6b-7dac-4bb3-b496-95d70839ff00-internal-tls-certs\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.152761 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd5faf6b-7dac-4bb3-b496-95d70839ff00-logs\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.152916 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd5faf6b-7dac-4bb3-b496-95d70839ff00-config-data\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.152957 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd5faf6b-7dac-4bb3-b496-95d70839ff00-combined-ca-bundle\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.153002 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wrxh\" (UniqueName: \"kubernetes.io/projected/fd5faf6b-7dac-4bb3-b496-95d70839ff00-kube-api-access-8wrxh\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.153362 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd5faf6b-7dac-4bb3-b496-95d70839ff00-config-data-custom\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.153510 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd5faf6b-7dac-4bb3-b496-95d70839ff00-public-tls-certs\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.254557 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd5faf6b-7dac-4bb3-b496-95d70839ff00-config-data-custom\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.254632 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd5faf6b-7dac-4bb3-b496-95d70839ff00-public-tls-certs\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.254675 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd5faf6b-7dac-4bb3-b496-95d70839ff00-internal-tls-certs\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.254719 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd5faf6b-7dac-4bb3-b496-95d70839ff00-logs\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.254762 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd5faf6b-7dac-4bb3-b496-95d70839ff00-config-data\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.254793 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd5faf6b-7dac-4bb3-b496-95d70839ff00-combined-ca-bundle\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.254951 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wrxh\" (UniqueName: \"kubernetes.io/projected/fd5faf6b-7dac-4bb3-b496-95d70839ff00-kube-api-access-8wrxh\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.255387 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd5faf6b-7dac-4bb3-b496-95d70839ff00-logs\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.263365 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd5faf6b-7dac-4bb3-b496-95d70839ff00-internal-tls-certs\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.264056 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd5faf6b-7dac-4bb3-b496-95d70839ff00-config-data-custom\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.265190 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd5faf6b-7dac-4bb3-b496-95d70839ff00-public-tls-certs\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.266751 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd5faf6b-7dac-4bb3-b496-95d70839ff00-combined-ca-bundle\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.268875 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd5faf6b-7dac-4bb3-b496-95d70839ff00-config-data\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.276481 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wrxh\" (UniqueName: \"kubernetes.io/projected/fd5faf6b-7dac-4bb3-b496-95d70839ff00-kube-api-access-8wrxh\") pod \"barbican-api-c9489b69d-msqgc\" (UID: \"fd5faf6b-7dac-4bb3-b496-95d70839ff00\") " pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.464337 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.780458 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6bc9d549b8-hdn6n" podUID="cfb252c0-ffb2-4d87-9c4e-b56cc2070d71" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.780537 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.781235 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"14da3471fa5fd4d6ec42e957717617fb2dea34019a855f12f66753cef241842c"} pod="openstack/horizon-6bc9d549b8-hdn6n" containerMessage="Container horizon failed startup probe, will be restarted" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.781271 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6bc9d549b8-hdn6n" podUID="cfb252c0-ffb2-4d87-9c4e-b56cc2070d71" containerName="horizon" containerID="cri-o://14da3471fa5fd4d6ec42e957717617fb2dea34019a855f12f66753cef241842c" gracePeriod=30 Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.950656 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-687657b496-tz596" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.950719 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-687657b496-tz596" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.951328 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"609e50be7f47738cb94497fefee796e278d503fdc50869f2e264123ccbb09fa6"} pod="openstack/horizon-687657b496-tz596" containerMessage="Container horizon failed startup probe, will be restarted" Feb 14 11:27:38 crc kubenswrapper[4904]: I0214 11:27:38.951360 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-687657b496-tz596" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon" containerID="cri-o://609e50be7f47738cb94497fefee796e278d503fdc50869f2e264123ccbb09fa6" gracePeriod=30 Feb 14 11:27:39 crc kubenswrapper[4904]: I0214 11:27:39.026688 4904 generic.go:334] "Generic (PLEG): container finished" podID="77818f86-c391-47b6-afff-5bf2831b9328" containerID="e07211442907acafa7321466c70ea387e33038cb4a3b210324aa5761de4e6ff1" exitCode=143 Feb 14 11:27:39 crc kubenswrapper[4904]: I0214 11:27:39.027535 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"77818f86-c391-47b6-afff-5bf2831b9328","Type":"ContainerDied","Data":"e07211442907acafa7321466c70ea387e33038cb4a3b210324aa5761de4e6ff1"} Feb 14 11:27:39 crc kubenswrapper[4904]: I0214 11:27:39.240657 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-58c9948449-g7lbx" podUID="4cc31d3d-1f06-48a1-9856-e1a997bfb11e" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.158:9696/\": dial tcp 10.217.0.158:9696: connect: connection refused" Feb 14 11:27:39 crc kubenswrapper[4904]: I0214 11:27:39.650893 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:27:39 crc kubenswrapper[4904]: I0214 11:27:39.818048 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c9489b69d-msqgc"] Feb 14 11:27:39 crc kubenswrapper[4904]: I0214 11:27:39.960568 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7856f6fd6f-4rhbv"] Feb 14 11:27:39 crc kubenswrapper[4904]: W0214 11:27:39.962114 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8261bc0e_8af3_4000_aa2d_de7153a235ca.slice/crio-896803bec86c53de2a33aeae142fcb113d74cbda85fc92b2752a89433a94557f WatchSource:0}: Error finding container 896803bec86c53de2a33aeae142fcb113d74cbda85fc92b2752a89433a94557f: Status 404 returned error can't find the container with id 896803bec86c53de2a33aeae142fcb113d74cbda85fc92b2752a89433a94557f Feb 14 11:27:40 crc kubenswrapper[4904]: I0214 11:27:40.049007 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-686dcf6685-s975t" event={"ID":"d096f44e-6631-414c-a590-e9e5f2f0cde2","Type":"ContainerStarted","Data":"f670bfa9170dbb19a80545b4cb744f8552190bf576d73c271accb006da9f9d30"} Feb 14 11:27:40 crc kubenswrapper[4904]: I0214 11:27:40.054415 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c9489b69d-msqgc" event={"ID":"fd5faf6b-7dac-4bb3-b496-95d70839ff00","Type":"ContainerStarted","Data":"8df23bb9a7262df2b0ac3aa2797c74393b37d71588a8d911b7eec85a56963ff8"} Feb 14 11:27:40 crc kubenswrapper[4904]: I0214 11:27:40.056301 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bccb523-3f10-41ba-aeba-0cd9675d20e2","Type":"ContainerStarted","Data":"a075ad508cae7d2e6bfb30368cf5d1ecfe6c32daa8e99efd51101a629719cb6b"} Feb 14 11:27:40 crc kubenswrapper[4904]: I0214 11:27:40.058770 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" event={"ID":"2f74e923-d64a-4161-9839-96a80dd93c2b","Type":"ContainerStarted","Data":"9ff1a712bb828c56f23903aae9568de53fa25d8f6ea238220dfcdc7c381d3637"} Feb 14 11:27:40 crc kubenswrapper[4904]: I0214 11:27:40.061282 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7856f6fd6f-4rhbv" event={"ID":"8261bc0e-8af3-4000-aa2d-de7153a235ca","Type":"ContainerStarted","Data":"896803bec86c53de2a33aeae142fcb113d74cbda85fc92b2752a89433a94557f"} Feb 14 11:27:40 crc kubenswrapper[4904]: I0214 11:27:40.941011 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 14 11:27:41 crc kubenswrapper[4904]: I0214 11:27:41.075305 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c9489b69d-msqgc" event={"ID":"fd5faf6b-7dac-4bb3-b496-95d70839ff00","Type":"ContainerStarted","Data":"1229dfc320db107f0fcf243f295bae22916baea77f4aa149380573733bbcf053"} Feb 14 11:27:41 crc kubenswrapper[4904]: I0214 11:27:41.075580 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c9489b69d-msqgc" event={"ID":"fd5faf6b-7dac-4bb3-b496-95d70839ff00","Type":"ContainerStarted","Data":"8180e14c73e180a104c7c20f445971689de90a13454f07933c431c3cc49aa700"} Feb 14 11:27:41 crc kubenswrapper[4904]: I0214 11:27:41.075811 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:41 crc kubenswrapper[4904]: I0214 11:27:41.075958 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:41 crc kubenswrapper[4904]: I0214 11:27:41.079874 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bccb523-3f10-41ba-aeba-0cd9675d20e2","Type":"ContainerStarted","Data":"0fc7805dae0f35163979dad6a3dc270a83092837b7fcf764383338be3e0d6038"} Feb 14 11:27:41 crc kubenswrapper[4904]: I0214 11:27:41.083285 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" event={"ID":"2f74e923-d64a-4161-9839-96a80dd93c2b","Type":"ContainerStarted","Data":"15fab531f64a8a56e86b92254bd910b56749cbeec67815ce2377dbf6069e5398"} Feb 14 11:27:41 crc kubenswrapper[4904]: I0214 11:27:41.089267 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7856f6fd6f-4rhbv" event={"ID":"8261bc0e-8af3-4000-aa2d-de7153a235ca","Type":"ContainerStarted","Data":"e558d516c9bda25ce3dcd46ff5a459c55ebc834c694c2ec3a65d6faa39d7ae6c"} Feb 14 11:27:41 crc kubenswrapper[4904]: I0214 11:27:41.089292 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7856f6fd6f-4rhbv" event={"ID":"8261bc0e-8af3-4000-aa2d-de7153a235ca","Type":"ContainerStarted","Data":"6dca0efa17c07d9bc98000ee3b144271ddef39b8b8067d88f46ba62bc1ac318e"} Feb 14 11:27:41 crc kubenswrapper[4904]: I0214 11:27:41.089750 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:27:41 crc kubenswrapper[4904]: I0214 11:27:41.095045 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-686dcf6685-s975t" event={"ID":"d096f44e-6631-414c-a590-e9e5f2f0cde2","Type":"ContainerStarted","Data":"72e50b4413d06c8601e8784c4f7693f48d4cd3e7bd5eb3e574c09217851c9529"} Feb 14 11:27:41 crc kubenswrapper[4904]: I0214 11:27:41.107043 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-c9489b69d-msqgc" podStartSLOduration=3.107024465 podStartE2EDuration="3.107024465s" podCreationTimestamp="2026-02-14 11:27:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:27:41.098280234 +0000 UTC m=+1051.911044895" watchObservedRunningTime="2026-02-14 11:27:41.107024465 +0000 UTC m=+1051.919789126" Feb 14 11:27:41 crc kubenswrapper[4904]: I0214 11:27:41.137201 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7856f6fd6f-4rhbv" podStartSLOduration=4.137181544 podStartE2EDuration="4.137181544s" podCreationTimestamp="2026-02-14 11:27:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:27:41.131786575 +0000 UTC m=+1051.944551236" watchObservedRunningTime="2026-02-14 11:27:41.137181544 +0000 UTC m=+1051.949946205" Feb 14 11:27:41 crc kubenswrapper[4904]: I0214 11:27:41.177113 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6dfd997684-xhxq7" podStartSLOduration=5.703430245 podStartE2EDuration="10.177092532s" podCreationTimestamp="2026-02-14 11:27:31 +0000 UTC" firstStartedPulling="2026-02-14 11:27:34.465265239 +0000 UTC m=+1045.278029900" lastFinishedPulling="2026-02-14 11:27:38.938927526 +0000 UTC m=+1049.751692187" observedRunningTime="2026-02-14 11:27:41.168813074 +0000 UTC m=+1051.981577735" watchObservedRunningTime="2026-02-14 11:27:41.177092532 +0000 UTC m=+1051.989857193" Feb 14 11:27:41 crc kubenswrapper[4904]: I0214 11:27:41.193242 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-686dcf6685-s975t" podStartSLOduration=5.709967725 podStartE2EDuration="10.193212165s" podCreationTimestamp="2026-02-14 11:27:31 +0000 UTC" firstStartedPulling="2026-02-14 11:27:34.456847738 +0000 UTC m=+1045.269612399" lastFinishedPulling="2026-02-14 11:27:38.940092178 +0000 UTC m=+1049.752856839" observedRunningTime="2026-02-14 11:27:41.190809839 +0000 UTC m=+1052.003574500" watchObservedRunningTime="2026-02-14 11:27:41.193212165 +0000 UTC m=+1052.005976826" Feb 14 11:27:42 crc kubenswrapper[4904]: I0214 11:27:42.104183 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bccb523-3f10-41ba-aeba-0cd9675d20e2","Type":"ContainerStarted","Data":"e19677930607ed94dd63e028a05e75517c49461ca3589ebee5ec18a9cecac409"} Feb 14 11:27:42 crc kubenswrapper[4904]: I0214 11:27:42.104487 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bccb523-3f10-41ba-aeba-0cd9675d20e2","Type":"ContainerStarted","Data":"9dbb902d96c46ece6c735688777a0dceda0d33dd9089c12d6ac517b70be69e1f"} Feb 14 11:27:42 crc kubenswrapper[4904]: I0214 11:27:42.242069 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:27:42 crc kubenswrapper[4904]: I0214 11:27:42.331536 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-zqqbz"] Feb 14 11:27:42 crc kubenswrapper[4904]: I0214 11:27:42.331767 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" podUID="5eaaa567-fbe1-4db5-822a-8b6f1ff773aa" containerName="dnsmasq-dns" containerID="cri-o://5478a0462c7ac487dd027e1b03f15769668ff98a095725635740e868c8f50bd1" gracePeriod=10 Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.125094 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.127170 4904 generic.go:334] "Generic (PLEG): container finished" podID="5eaaa567-fbe1-4db5-822a-8b6f1ff773aa" containerID="5478a0462c7ac487dd027e1b03f15769668ff98a095725635740e868c8f50bd1" exitCode=0 Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.128072 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" event={"ID":"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa","Type":"ContainerDied","Data":"5478a0462c7ac487dd027e1b03f15769668ff98a095725635740e868c8f50bd1"} Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.128097 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" event={"ID":"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa","Type":"ContainerDied","Data":"fc9619619b133fd11c3cf64650614957dcf50fa8a8e414f35f7fbec0544e42c8"} Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.128116 4904 scope.go:117] "RemoveContainer" containerID="5478a0462c7ac487dd027e1b03f15769668ff98a095725635740e868c8f50bd1" Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.171267 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjmgp\" (UniqueName: \"kubernetes.io/projected/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-kube-api-access-rjmgp\") pod \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.171996 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-ovsdbserver-nb\") pod \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.172170 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-config\") pod \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.172277 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-dns-svc\") pod \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.172404 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-ovsdbserver-sb\") pod \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.172588 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-dns-swift-storage-0\") pod \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\" (UID: \"5eaaa567-fbe1-4db5-822a-8b6f1ff773aa\") " Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.172754 4904 scope.go:117] "RemoveContainer" containerID="7d5e8e8081b7b6b9665f4e4356f8c5968e9dceee7028186d474ec75aa35a72a8" Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.206966 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-kube-api-access-rjmgp" (OuterVolumeSpecName: "kube-api-access-rjmgp") pod "5eaaa567-fbe1-4db5-822a-8b6f1ff773aa" (UID: "5eaaa567-fbe1-4db5-822a-8b6f1ff773aa"). InnerVolumeSpecName "kube-api-access-rjmgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.269708 4904 scope.go:117] "RemoveContainer" containerID="5478a0462c7ac487dd027e1b03f15769668ff98a095725635740e868c8f50bd1" Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.278205 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjmgp\" (UniqueName: \"kubernetes.io/projected/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-kube-api-access-rjmgp\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:43 crc kubenswrapper[4904]: E0214 11:27:43.284024 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5478a0462c7ac487dd027e1b03f15769668ff98a095725635740e868c8f50bd1\": container with ID starting with 5478a0462c7ac487dd027e1b03f15769668ff98a095725635740e868c8f50bd1 not found: ID does not exist" containerID="5478a0462c7ac487dd027e1b03f15769668ff98a095725635740e868c8f50bd1" Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.284068 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5478a0462c7ac487dd027e1b03f15769668ff98a095725635740e868c8f50bd1"} err="failed to get container status \"5478a0462c7ac487dd027e1b03f15769668ff98a095725635740e868c8f50bd1\": rpc error: code = NotFound desc = could not find container \"5478a0462c7ac487dd027e1b03f15769668ff98a095725635740e868c8f50bd1\": container with ID starting with 5478a0462c7ac487dd027e1b03f15769668ff98a095725635740e868c8f50bd1 not found: ID does not exist" Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.284093 4904 scope.go:117] "RemoveContainer" containerID="7d5e8e8081b7b6b9665f4e4356f8c5968e9dceee7028186d474ec75aa35a72a8" Feb 14 11:27:43 crc kubenswrapper[4904]: E0214 11:27:43.293021 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d5e8e8081b7b6b9665f4e4356f8c5968e9dceee7028186d474ec75aa35a72a8\": container with ID starting with 7d5e8e8081b7b6b9665f4e4356f8c5968e9dceee7028186d474ec75aa35a72a8 not found: ID does not exist" containerID="7d5e8e8081b7b6b9665f4e4356f8c5968e9dceee7028186d474ec75aa35a72a8" Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.293060 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d5e8e8081b7b6b9665f4e4356f8c5968e9dceee7028186d474ec75aa35a72a8"} err="failed to get container status \"7d5e8e8081b7b6b9665f4e4356f8c5968e9dceee7028186d474ec75aa35a72a8\": rpc error: code = NotFound desc = could not find container \"7d5e8e8081b7b6b9665f4e4356f8c5968e9dceee7028186d474ec75aa35a72a8\": container with ID starting with 7d5e8e8081b7b6b9665f4e4356f8c5968e9dceee7028186d474ec75aa35a72a8 not found: ID does not exist" Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.354558 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5eaaa567-fbe1-4db5-822a-8b6f1ff773aa" (UID: "5eaaa567-fbe1-4db5-822a-8b6f1ff773aa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.357388 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5eaaa567-fbe1-4db5-822a-8b6f1ff773aa" (UID: "5eaaa567-fbe1-4db5-822a-8b6f1ff773aa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.374656 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-config" (OuterVolumeSpecName: "config") pod "5eaaa567-fbe1-4db5-822a-8b6f1ff773aa" (UID: "5eaaa567-fbe1-4db5-822a-8b6f1ff773aa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.381076 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.381111 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.381120 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.384382 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5eaaa567-fbe1-4db5-822a-8b6f1ff773aa" (UID: "5eaaa567-fbe1-4db5-822a-8b6f1ff773aa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.412257 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5eaaa567-fbe1-4db5-822a-8b6f1ff773aa" (UID: "5eaaa567-fbe1-4db5-822a-8b6f1ff773aa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.483189 4904 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:43 crc kubenswrapper[4904]: I0214 11:27:43.483224 4904 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.167239 4904 generic.go:334] "Generic (PLEG): container finished" podID="4cc31d3d-1f06-48a1-9856-e1a997bfb11e" containerID="df39d52890d4f62c453dfa0452f433a477f0e1ede96fc0bb86defbf9ea1c50a1" exitCode=0 Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.167602 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58c9948449-g7lbx" event={"ID":"4cc31d3d-1f06-48a1-9856-e1a997bfb11e","Type":"ContainerDied","Data":"df39d52890d4f62c453dfa0452f433a477f0e1ede96fc0bb86defbf9ea1c50a1"} Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.172286 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bccb523-3f10-41ba-aeba-0cd9675d20e2","Type":"ContainerStarted","Data":"aff6f3d23a47f62fe7f814538d959f2633736b7949ea773881880c1e95b1d205"} Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.172374 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.174178 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-zqqbz" Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.253054 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.982966064 podStartE2EDuration="8.253036515s" podCreationTimestamp="2026-02-14 11:27:36 +0000 UTC" firstStartedPulling="2026-02-14 11:27:39.739012287 +0000 UTC m=+1050.551776948" lastFinishedPulling="2026-02-14 11:27:43.009082738 +0000 UTC m=+1053.821847399" observedRunningTime="2026-02-14 11:27:44.209609721 +0000 UTC m=+1055.022374382" watchObservedRunningTime="2026-02-14 11:27:44.253036515 +0000 UTC m=+1055.065801176" Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.257816 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-zqqbz"] Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.274294 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-zqqbz"] Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.458666 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.507028 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-combined-ca-bundle\") pod \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.507172 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-public-tls-certs\") pod \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.507204 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-internal-tls-certs\") pod \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.507239 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-config\") pod \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.507275 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-httpd-config\") pod \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.507330 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnskp\" (UniqueName: \"kubernetes.io/projected/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-kube-api-access-mnskp\") pod \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.507369 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-ovndb-tls-certs\") pod \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\" (UID: \"4cc31d3d-1f06-48a1-9856-e1a997bfb11e\") " Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.540066 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-kube-api-access-mnskp" (OuterVolumeSpecName: "kube-api-access-mnskp") pod "4cc31d3d-1f06-48a1-9856-e1a997bfb11e" (UID: "4cc31d3d-1f06-48a1-9856-e1a997bfb11e"). InnerVolumeSpecName "kube-api-access-mnskp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.566742 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "4cc31d3d-1f06-48a1-9856-e1a997bfb11e" (UID: "4cc31d3d-1f06-48a1-9856-e1a997bfb11e"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.601709 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-config" (OuterVolumeSpecName: "config") pod "4cc31d3d-1f06-48a1-9856-e1a997bfb11e" (UID: "4cc31d3d-1f06-48a1-9856-e1a997bfb11e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.616636 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.616667 4904 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.616676 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnskp\" (UniqueName: \"kubernetes.io/projected/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-kube-api-access-mnskp\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.643941 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4cc31d3d-1f06-48a1-9856-e1a997bfb11e" (UID: "4cc31d3d-1f06-48a1-9856-e1a997bfb11e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.694914 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4cc31d3d-1f06-48a1-9856-e1a997bfb11e" (UID: "4cc31d3d-1f06-48a1-9856-e1a997bfb11e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.721023 4904 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.721051 4904 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.740329 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4cc31d3d-1f06-48a1-9856-e1a997bfb11e" (UID: "4cc31d3d-1f06-48a1-9856-e1a997bfb11e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.757041 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "4cc31d3d-1f06-48a1-9856-e1a997bfb11e" (UID: "4cc31d3d-1f06-48a1-9856-e1a997bfb11e"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.823111 4904 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:44 crc kubenswrapper[4904]: I0214 11:27:44.823144 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cc31d3d-1f06-48a1-9856-e1a997bfb11e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:45 crc kubenswrapper[4904]: I0214 11:27:45.183968 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58c9948449-g7lbx" Feb 14 11:27:45 crc kubenswrapper[4904]: I0214 11:27:45.183967 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58c9948449-g7lbx" event={"ID":"4cc31d3d-1f06-48a1-9856-e1a997bfb11e","Type":"ContainerDied","Data":"abc724f0405ac23e7f36a76f074c375906ad3ae83471a52035a6ca5949bbeabb"} Feb 14 11:27:45 crc kubenswrapper[4904]: I0214 11:27:45.184036 4904 scope.go:117] "RemoveContainer" containerID="eea579d5897474d21fdec00430fbfcb39d8f2d878d90ca0eb9305bd3657fc893" Feb 14 11:27:45 crc kubenswrapper[4904]: I0214 11:27:45.207884 4904 scope.go:117] "RemoveContainer" containerID="df39d52890d4f62c453dfa0452f433a477f0e1ede96fc0bb86defbf9ea1c50a1" Feb 14 11:27:45 crc kubenswrapper[4904]: I0214 11:27:45.230410 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-58c9948449-g7lbx"] Feb 14 11:27:45 crc kubenswrapper[4904]: I0214 11:27:45.240294 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-58c9948449-g7lbx"] Feb 14 11:27:45 crc kubenswrapper[4904]: I0214 11:27:45.845965 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cc31d3d-1f06-48a1-9856-e1a997bfb11e" path="/var/lib/kubelet/pods/4cc31d3d-1f06-48a1-9856-e1a997bfb11e/volumes" Feb 14 11:27:45 crc kubenswrapper[4904]: I0214 11:27:45.847008 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5eaaa567-fbe1-4db5-822a-8b6f1ff773aa" path="/var/lib/kubelet/pods/5eaaa567-fbe1-4db5-822a-8b6f1ff773aa/volumes" Feb 14 11:27:45 crc kubenswrapper[4904]: I0214 11:27:45.946270 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:46 crc kubenswrapper[4904]: I0214 11:27:46.100791 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:27:46 crc kubenswrapper[4904]: I0214 11:27:46.336804 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 14 11:27:46 crc kubenswrapper[4904]: I0214 11:27:46.418013 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 14 11:27:47 crc kubenswrapper[4904]: I0214 11:27:47.200047 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="56cfc647-7ed2-4649-979b-8a3cd60bd44f" containerName="cinder-scheduler" containerID="cri-o://22715d0f5fd42bf2706df319554222c8e0581c422c1f7dcaa7186748f5c37c3e" gracePeriod=30 Feb 14 11:27:47 crc kubenswrapper[4904]: I0214 11:27:47.200137 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="56cfc647-7ed2-4649-979b-8a3cd60bd44f" containerName="probe" containerID="cri-o://77445f189d44c809b28038a9734762f33fa07206f66b5250894d21888fd291ac" gracePeriod=30 Feb 14 11:27:48 crc kubenswrapper[4904]: I0214 11:27:48.218185 4904 generic.go:334] "Generic (PLEG): container finished" podID="56cfc647-7ed2-4649-979b-8a3cd60bd44f" containerID="77445f189d44c809b28038a9734762f33fa07206f66b5250894d21888fd291ac" exitCode=0 Feb 14 11:27:48 crc kubenswrapper[4904]: I0214 11:27:48.218260 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"56cfc647-7ed2-4649-979b-8a3cd60bd44f","Type":"ContainerDied","Data":"77445f189d44c809b28038a9734762f33fa07206f66b5250894d21888fd291ac"} Feb 14 11:27:49 crc kubenswrapper[4904]: I0214 11:27:49.721767 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.241726 4904 generic.go:334] "Generic (PLEG): container finished" podID="56cfc647-7ed2-4649-979b-8a3cd60bd44f" containerID="22715d0f5fd42bf2706df319554222c8e0581c422c1f7dcaa7186748f5c37c3e" exitCode=0 Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.241975 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"56cfc647-7ed2-4649-979b-8a3cd60bd44f","Type":"ContainerDied","Data":"22715d0f5fd42bf2706df319554222c8e0581c422c1f7dcaa7186748f5c37c3e"} Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.345604 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.434076 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-combined-ca-bundle\") pod \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.434366 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-config-data\") pod \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.434407 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-config-data-custom\") pod \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.435686 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twqp4\" (UniqueName: \"kubernetes.io/projected/56cfc647-7ed2-4649-979b-8a3cd60bd44f-kube-api-access-twqp4\") pod \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.435722 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/56cfc647-7ed2-4649-979b-8a3cd60bd44f-etc-machine-id\") pod \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.435767 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-scripts\") pod \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\" (UID: \"56cfc647-7ed2-4649-979b-8a3cd60bd44f\") " Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.442961 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/56cfc647-7ed2-4649-979b-8a3cd60bd44f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "56cfc647-7ed2-4649-979b-8a3cd60bd44f" (UID: "56cfc647-7ed2-4649-979b-8a3cd60bd44f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.449402 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "56cfc647-7ed2-4649-979b-8a3cd60bd44f" (UID: "56cfc647-7ed2-4649-979b-8a3cd60bd44f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.456047 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56cfc647-7ed2-4649-979b-8a3cd60bd44f-kube-api-access-twqp4" (OuterVolumeSpecName: "kube-api-access-twqp4") pod "56cfc647-7ed2-4649-979b-8a3cd60bd44f" (UID: "56cfc647-7ed2-4649-979b-8a3cd60bd44f"). InnerVolumeSpecName "kube-api-access-twqp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.475027 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-scripts" (OuterVolumeSpecName: "scripts") pod "56cfc647-7ed2-4649-979b-8a3cd60bd44f" (UID: "56cfc647-7ed2-4649-979b-8a3cd60bd44f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.537549 4904 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.537575 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twqp4\" (UniqueName: \"kubernetes.io/projected/56cfc647-7ed2-4649-979b-8a3cd60bd44f-kube-api-access-twqp4\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.537587 4904 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/56cfc647-7ed2-4649-979b-8a3cd60bd44f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.537596 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.553973 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56cfc647-7ed2-4649-979b-8a3cd60bd44f" (UID: "56cfc647-7ed2-4649-979b-8a3cd60bd44f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.648932 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.741231 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-config-data" (OuterVolumeSpecName: "config-data") pod "56cfc647-7ed2-4649-979b-8a3cd60bd44f" (UID: "56cfc647-7ed2-4649-979b-8a3cd60bd44f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:27:50 crc kubenswrapper[4904]: I0214 11:27:50.755127 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56cfc647-7ed2-4649-979b-8a3cd60bd44f-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.250789 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"56cfc647-7ed2-4649-979b-8a3cd60bd44f","Type":"ContainerDied","Data":"d6370b6c19b5bc0cae9cf3324da385a7898ed64417ddbe121d1235f21bfd6886"} Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.250852 4904 scope.go:117] "RemoveContainer" containerID="77445f189d44c809b28038a9734762f33fa07206f66b5250894d21888fd291ac" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.250870 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.329177 4904 scope.go:117] "RemoveContainer" containerID="22715d0f5fd42bf2706df319554222c8e0581c422c1f7dcaa7186748f5c37c3e" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.343016 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.378818 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.402714 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 14 11:27:51 crc kubenswrapper[4904]: E0214 11:27:51.403210 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56cfc647-7ed2-4649-979b-8a3cd60bd44f" containerName="probe" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.403226 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="56cfc647-7ed2-4649-979b-8a3cd60bd44f" containerName="probe" Feb 14 11:27:51 crc kubenswrapper[4904]: E0214 11:27:51.403236 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eaaa567-fbe1-4db5-822a-8b6f1ff773aa" containerName="init" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.403243 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eaaa567-fbe1-4db5-822a-8b6f1ff773aa" containerName="init" Feb 14 11:27:51 crc kubenswrapper[4904]: E0214 11:27:51.403252 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc31d3d-1f06-48a1-9856-e1a997bfb11e" containerName="neutron-api" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.403259 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc31d3d-1f06-48a1-9856-e1a997bfb11e" containerName="neutron-api" Feb 14 11:27:51 crc kubenswrapper[4904]: E0214 11:27:51.403292 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56cfc647-7ed2-4649-979b-8a3cd60bd44f" containerName="cinder-scheduler" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.403299 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="56cfc647-7ed2-4649-979b-8a3cd60bd44f" containerName="cinder-scheduler" Feb 14 11:27:51 crc kubenswrapper[4904]: E0214 11:27:51.403314 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eaaa567-fbe1-4db5-822a-8b6f1ff773aa" containerName="dnsmasq-dns" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.403321 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eaaa567-fbe1-4db5-822a-8b6f1ff773aa" containerName="dnsmasq-dns" Feb 14 11:27:51 crc kubenswrapper[4904]: E0214 11:27:51.403336 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc31d3d-1f06-48a1-9856-e1a997bfb11e" containerName="neutron-httpd" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.403342 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc31d3d-1f06-48a1-9856-e1a997bfb11e" containerName="neutron-httpd" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.403508 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="5eaaa567-fbe1-4db5-822a-8b6f1ff773aa" containerName="dnsmasq-dns" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.403524 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="56cfc647-7ed2-4649-979b-8a3cd60bd44f" containerName="cinder-scheduler" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.403533 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="56cfc647-7ed2-4649-979b-8a3cd60bd44f" containerName="probe" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.403543 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cc31d3d-1f06-48a1-9856-e1a997bfb11e" containerName="neutron-api" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.403558 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cc31d3d-1f06-48a1-9856-e1a997bfb11e" containerName="neutron-httpd" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.404654 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.413019 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.417987 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.472956 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b66062f0-a15b-41cb-b68f-753186581ae2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b66062f0-a15b-41cb-b68f-753186581ae2\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.473006 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b66062f0-a15b-41cb-b68f-753186581ae2-scripts\") pod \"cinder-scheduler-0\" (UID: \"b66062f0-a15b-41cb-b68f-753186581ae2\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.473078 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b66062f0-a15b-41cb-b68f-753186581ae2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b66062f0-a15b-41cb-b68f-753186581ae2\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.473136 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b66062f0-a15b-41cb-b68f-753186581ae2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b66062f0-a15b-41cb-b68f-753186581ae2\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.473154 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh79h\" (UniqueName: \"kubernetes.io/projected/b66062f0-a15b-41cb-b68f-753186581ae2-kube-api-access-kh79h\") pod \"cinder-scheduler-0\" (UID: \"b66062f0-a15b-41cb-b68f-753186581ae2\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.473196 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b66062f0-a15b-41cb-b68f-753186581ae2-config-data\") pod \"cinder-scheduler-0\" (UID: \"b66062f0-a15b-41cb-b68f-753186581ae2\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.574775 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b66062f0-a15b-41cb-b68f-753186581ae2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b66062f0-a15b-41cb-b68f-753186581ae2\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.574865 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b66062f0-a15b-41cb-b68f-753186581ae2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b66062f0-a15b-41cb-b68f-753186581ae2\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.574888 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh79h\" (UniqueName: \"kubernetes.io/projected/b66062f0-a15b-41cb-b68f-753186581ae2-kube-api-access-kh79h\") pod \"cinder-scheduler-0\" (UID: \"b66062f0-a15b-41cb-b68f-753186581ae2\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.574907 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b66062f0-a15b-41cb-b68f-753186581ae2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b66062f0-a15b-41cb-b68f-753186581ae2\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.574941 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b66062f0-a15b-41cb-b68f-753186581ae2-config-data\") pod \"cinder-scheduler-0\" (UID: \"b66062f0-a15b-41cb-b68f-753186581ae2\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.575000 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b66062f0-a15b-41cb-b68f-753186581ae2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b66062f0-a15b-41cb-b68f-753186581ae2\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.575026 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b66062f0-a15b-41cb-b68f-753186581ae2-scripts\") pod \"cinder-scheduler-0\" (UID: \"b66062f0-a15b-41cb-b68f-753186581ae2\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.578391 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b66062f0-a15b-41cb-b68f-753186581ae2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b66062f0-a15b-41cb-b68f-753186581ae2\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.581147 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b66062f0-a15b-41cb-b68f-753186581ae2-scripts\") pod \"cinder-scheduler-0\" (UID: \"b66062f0-a15b-41cb-b68f-753186581ae2\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.584410 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b66062f0-a15b-41cb-b68f-753186581ae2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b66062f0-a15b-41cb-b68f-753186581ae2\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.584740 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b66062f0-a15b-41cb-b68f-753186581ae2-config-data\") pod \"cinder-scheduler-0\" (UID: \"b66062f0-a15b-41cb-b68f-753186581ae2\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.591991 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh79h\" (UniqueName: \"kubernetes.io/projected/b66062f0-a15b-41cb-b68f-753186581ae2-kube-api-access-kh79h\") pod \"cinder-scheduler-0\" (UID: \"b66062f0-a15b-41cb-b68f-753186581ae2\") " pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.746982 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.854061 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56cfc647-7ed2-4649-979b-8a3cd60bd44f" path="/var/lib/kubelet/pods/56cfc647-7ed2-4649-979b-8a3cd60bd44f/volumes" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.921935 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:51 crc kubenswrapper[4904]: I0214 11:27:51.945203 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-bb76f8864-hs52l" Feb 14 11:27:52 crc kubenswrapper[4904]: I0214 11:27:52.430130 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 14 11:27:52 crc kubenswrapper[4904]: I0214 11:27:52.477018 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-c9489b69d-msqgc" podUID="fd5faf6b-7dac-4bb3-b496-95d70839ff00" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.170:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 14 11:27:52 crc kubenswrapper[4904]: I0214 11:27:52.871335 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:53 crc kubenswrapper[4904]: I0214 11:27:53.302396 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b66062f0-a15b-41cb-b68f-753186581ae2","Type":"ContainerStarted","Data":"9a2c0e3893f34e1d70dfccf16f771c9b380c57ecc880860ea3d4023de2c02a6b"} Feb 14 11:27:53 crc kubenswrapper[4904]: I0214 11:27:53.302434 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b66062f0-a15b-41cb-b68f-753186581ae2","Type":"ContainerStarted","Data":"7c71df0b29ac3b82de12b6fd8d0916b762b8e96f657b5f6fdd22e8ee11f1d06f"} Feb 14 11:27:53 crc kubenswrapper[4904]: I0214 11:27:53.471062 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-c9489b69d-msqgc" podUID="fd5faf6b-7dac-4bb3-b496-95d70839ff00" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.170:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 14 11:27:53 crc kubenswrapper[4904]: I0214 11:27:53.956288 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7b79cd46c9-g4bx8" Feb 14 11:27:55 crc kubenswrapper[4904]: I0214 11:27:55.319670 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b66062f0-a15b-41cb-b68f-753186581ae2","Type":"ContainerStarted","Data":"77d63bb097dbc87ab1aeb3e1d0d02211542140cb8e313e2a27bdbbb0a0df691b"} Feb 14 11:27:55 crc kubenswrapper[4904]: I0214 11:27:55.339320 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.3393017369999995 podStartE2EDuration="4.339301737s" podCreationTimestamp="2026-02-14 11:27:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:27:55.334360541 +0000 UTC m=+1066.147125202" watchObservedRunningTime="2026-02-14 11:27:55.339301737 +0000 UTC m=+1066.152066398" Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.294195 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="77818f86-c391-47b6-afff-5bf2831b9328" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.163:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.437048 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.438392 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.445116 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.445865 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-hs5ss" Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.447712 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.450034 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.500598 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twdmr\" (UniqueName: \"kubernetes.io/projected/ab9aab3c-a75c-4f0a-9125-623bb22314d0-kube-api-access-twdmr\") pod \"openstackclient\" (UID: \"ab9aab3c-a75c-4f0a-9125-623bb22314d0\") " pod="openstack/openstackclient" Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.501015 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab9aab3c-a75c-4f0a-9125-623bb22314d0-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ab9aab3c-a75c-4f0a-9125-623bb22314d0\") " pod="openstack/openstackclient" Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.501066 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ab9aab3c-a75c-4f0a-9125-623bb22314d0-openstack-config\") pod \"openstackclient\" (UID: \"ab9aab3c-a75c-4f0a-9125-623bb22314d0\") " pod="openstack/openstackclient" Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.501094 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ab9aab3c-a75c-4f0a-9125-623bb22314d0-openstack-config-secret\") pod \"openstackclient\" (UID: \"ab9aab3c-a75c-4f0a-9125-623bb22314d0\") " pod="openstack/openstackclient" Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.602272 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ab9aab3c-a75c-4f0a-9125-623bb22314d0-openstack-config-secret\") pod \"openstackclient\" (UID: \"ab9aab3c-a75c-4f0a-9125-623bb22314d0\") " pod="openstack/openstackclient" Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.602341 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twdmr\" (UniqueName: \"kubernetes.io/projected/ab9aab3c-a75c-4f0a-9125-623bb22314d0-kube-api-access-twdmr\") pod \"openstackclient\" (UID: \"ab9aab3c-a75c-4f0a-9125-623bb22314d0\") " pod="openstack/openstackclient" Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.602432 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab9aab3c-a75c-4f0a-9125-623bb22314d0-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ab9aab3c-a75c-4f0a-9125-623bb22314d0\") " pod="openstack/openstackclient" Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.602481 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ab9aab3c-a75c-4f0a-9125-623bb22314d0-openstack-config\") pod \"openstackclient\" (UID: \"ab9aab3c-a75c-4f0a-9125-623bb22314d0\") " pod="openstack/openstackclient" Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.603331 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ab9aab3c-a75c-4f0a-9125-623bb22314d0-openstack-config\") pod \"openstackclient\" (UID: \"ab9aab3c-a75c-4f0a-9125-623bb22314d0\") " pod="openstack/openstackclient" Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.609544 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab9aab3c-a75c-4f0a-9125-623bb22314d0-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ab9aab3c-a75c-4f0a-9125-623bb22314d0\") " pod="openstack/openstackclient" Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.614252 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ab9aab3c-a75c-4f0a-9125-623bb22314d0-openstack-config-secret\") pod \"openstackclient\" (UID: \"ab9aab3c-a75c-4f0a-9125-623bb22314d0\") " pod="openstack/openstackclient" Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.625041 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twdmr\" (UniqueName: \"kubernetes.io/projected/ab9aab3c-a75c-4f0a-9125-623bb22314d0-kube-api-access-twdmr\") pod \"openstackclient\" (UID: \"ab9aab3c-a75c-4f0a-9125-623bb22314d0\") " pod="openstack/openstackclient" Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.748045 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 14 11:27:56 crc kubenswrapper[4904]: I0214 11:27:56.755535 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 14 11:27:57 crc kubenswrapper[4904]: I0214 11:27:57.132824 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-c9489b69d-msqgc" Feb 14 11:27:57 crc kubenswrapper[4904]: I0214 11:27:57.213361 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-797ff5c494-4sfjf"] Feb 14 11:27:57 crc kubenswrapper[4904]: I0214 11:27:57.213631 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-797ff5c494-4sfjf" podUID="b12a8202-af0b-403e-9767-7cbaeb661cd5" containerName="barbican-api-log" containerID="cri-o://c8a3039de6035c530e1e87fcfbe0392cf888fdba457441ecf5e0475fb63ac5c7" gracePeriod=30 Feb 14 11:27:57 crc kubenswrapper[4904]: I0214 11:27:57.214119 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-797ff5c494-4sfjf" podUID="b12a8202-af0b-403e-9767-7cbaeb661cd5" containerName="barbican-api" containerID="cri-o://9dfd9c5085ab7919e13381193642df7de78c9d7afa7829b6627ba596f57c0a72" gracePeriod=30 Feb 14 11:27:57 crc kubenswrapper[4904]: I0214 11:27:57.313801 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 14 11:27:57 crc kubenswrapper[4904]: I0214 11:27:57.346822 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ab9aab3c-a75c-4f0a-9125-623bb22314d0","Type":"ContainerStarted","Data":"30a13be2c3ee0ed2d44719abe096548ce16833538490fc42dff532cf1ff7eaf0"} Feb 14 11:27:58 crc kubenswrapper[4904]: I0214 11:27:58.363513 4904 generic.go:334] "Generic (PLEG): container finished" podID="b12a8202-af0b-403e-9767-7cbaeb661cd5" containerID="c8a3039de6035c530e1e87fcfbe0392cf888fdba457441ecf5e0475fb63ac5c7" exitCode=143 Feb 14 11:27:58 crc kubenswrapper[4904]: I0214 11:27:58.363564 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-797ff5c494-4sfjf" event={"ID":"b12a8202-af0b-403e-9767-7cbaeb661cd5","Type":"ContainerDied","Data":"c8a3039de6035c530e1e87fcfbe0392cf888fdba457441ecf5e0475fb63ac5c7"} Feb 14 11:28:00 crc kubenswrapper[4904]: I0214 11:28:00.670195 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-797ff5c494-4sfjf" podUID="b12a8202-af0b-403e-9767-7cbaeb661cd5" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.167:9311/healthcheck\": read tcp 10.217.0.2:38930->10.217.0.167:9311: read: connection reset by peer" Feb 14 11:28:00 crc kubenswrapper[4904]: I0214 11:28:00.670194 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-797ff5c494-4sfjf" podUID="b12a8202-af0b-403e-9767-7cbaeb661cd5" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.167:9311/healthcheck\": read tcp 10.217.0.2:38926->10.217.0.167:9311: read: connection reset by peer" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.296062 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.399263 4904 generic.go:334] "Generic (PLEG): container finished" podID="b12a8202-af0b-403e-9767-7cbaeb661cd5" containerID="9dfd9c5085ab7919e13381193642df7de78c9d7afa7829b6627ba596f57c0a72" exitCode=0 Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.399303 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-797ff5c494-4sfjf" event={"ID":"b12a8202-af0b-403e-9767-7cbaeb661cd5","Type":"ContainerDied","Data":"9dfd9c5085ab7919e13381193642df7de78c9d7afa7829b6627ba596f57c0a72"} Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.399328 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-797ff5c494-4sfjf" event={"ID":"b12a8202-af0b-403e-9767-7cbaeb661cd5","Type":"ContainerDied","Data":"fd08f31feb180fc1a7ce12ed8a4dae1b50ad042c55310211dfd66ab94d9a0eee"} Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.399344 4904 scope.go:117] "RemoveContainer" containerID="9dfd9c5085ab7919e13381193642df7de78c9d7afa7829b6627ba596f57c0a72" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.399460 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-797ff5c494-4sfjf" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.416109 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b12a8202-af0b-403e-9767-7cbaeb661cd5-config-data-custom\") pod \"b12a8202-af0b-403e-9767-7cbaeb661cd5\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.416246 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-974wd\" (UniqueName: \"kubernetes.io/projected/b12a8202-af0b-403e-9767-7cbaeb661cd5-kube-api-access-974wd\") pod \"b12a8202-af0b-403e-9767-7cbaeb661cd5\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.416290 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b12a8202-af0b-403e-9767-7cbaeb661cd5-combined-ca-bundle\") pod \"b12a8202-af0b-403e-9767-7cbaeb661cd5\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.416332 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b12a8202-af0b-403e-9767-7cbaeb661cd5-config-data\") pod \"b12a8202-af0b-403e-9767-7cbaeb661cd5\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.416359 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b12a8202-af0b-403e-9767-7cbaeb661cd5-logs\") pod \"b12a8202-af0b-403e-9767-7cbaeb661cd5\" (UID: \"b12a8202-af0b-403e-9767-7cbaeb661cd5\") " Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.417351 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b12a8202-af0b-403e-9767-7cbaeb661cd5-logs" (OuterVolumeSpecName: "logs") pod "b12a8202-af0b-403e-9767-7cbaeb661cd5" (UID: "b12a8202-af0b-403e-9767-7cbaeb661cd5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.442779 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b12a8202-af0b-403e-9767-7cbaeb661cd5-kube-api-access-974wd" (OuterVolumeSpecName: "kube-api-access-974wd") pod "b12a8202-af0b-403e-9767-7cbaeb661cd5" (UID: "b12a8202-af0b-403e-9767-7cbaeb661cd5"). InnerVolumeSpecName "kube-api-access-974wd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.444057 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b12a8202-af0b-403e-9767-7cbaeb661cd5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b12a8202-af0b-403e-9767-7cbaeb661cd5" (UID: "b12a8202-af0b-403e-9767-7cbaeb661cd5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.449914 4904 scope.go:117] "RemoveContainer" containerID="c8a3039de6035c530e1e87fcfbe0392cf888fdba457441ecf5e0475fb63ac5c7" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.461603 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b12a8202-af0b-403e-9767-7cbaeb661cd5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b12a8202-af0b-403e-9767-7cbaeb661cd5" (UID: "b12a8202-af0b-403e-9767-7cbaeb661cd5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.525607 4904 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b12a8202-af0b-403e-9767-7cbaeb661cd5-logs\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.525866 4904 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b12a8202-af0b-403e-9767-7cbaeb661cd5-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.525963 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-974wd\" (UniqueName: \"kubernetes.io/projected/b12a8202-af0b-403e-9767-7cbaeb661cd5-kube-api-access-974wd\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.526046 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b12a8202-af0b-403e-9767-7cbaeb661cd5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.537886 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b12a8202-af0b-403e-9767-7cbaeb661cd5-config-data" (OuterVolumeSpecName: "config-data") pod "b12a8202-af0b-403e-9767-7cbaeb661cd5" (UID: "b12a8202-af0b-403e-9767-7cbaeb661cd5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.540425 4904 scope.go:117] "RemoveContainer" containerID="9dfd9c5085ab7919e13381193642df7de78c9d7afa7829b6627ba596f57c0a72" Feb 14 11:28:01 crc kubenswrapper[4904]: E0214 11:28:01.540794 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dfd9c5085ab7919e13381193642df7de78c9d7afa7829b6627ba596f57c0a72\": container with ID starting with 9dfd9c5085ab7919e13381193642df7de78c9d7afa7829b6627ba596f57c0a72 not found: ID does not exist" containerID="9dfd9c5085ab7919e13381193642df7de78c9d7afa7829b6627ba596f57c0a72" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.540967 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dfd9c5085ab7919e13381193642df7de78c9d7afa7829b6627ba596f57c0a72"} err="failed to get container status \"9dfd9c5085ab7919e13381193642df7de78c9d7afa7829b6627ba596f57c0a72\": rpc error: code = NotFound desc = could not find container \"9dfd9c5085ab7919e13381193642df7de78c9d7afa7829b6627ba596f57c0a72\": container with ID starting with 9dfd9c5085ab7919e13381193642df7de78c9d7afa7829b6627ba596f57c0a72 not found: ID does not exist" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.541000 4904 scope.go:117] "RemoveContainer" containerID="c8a3039de6035c530e1e87fcfbe0392cf888fdba457441ecf5e0475fb63ac5c7" Feb 14 11:28:01 crc kubenswrapper[4904]: E0214 11:28:01.541610 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8a3039de6035c530e1e87fcfbe0392cf888fdba457441ecf5e0475fb63ac5c7\": container with ID starting with c8a3039de6035c530e1e87fcfbe0392cf888fdba457441ecf5e0475fb63ac5c7 not found: ID does not exist" containerID="c8a3039de6035c530e1e87fcfbe0392cf888fdba457441ecf5e0475fb63ac5c7" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.541647 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8a3039de6035c530e1e87fcfbe0392cf888fdba457441ecf5e0475fb63ac5c7"} err="failed to get container status \"c8a3039de6035c530e1e87fcfbe0392cf888fdba457441ecf5e0475fb63ac5c7\": rpc error: code = NotFound desc = could not find container \"c8a3039de6035c530e1e87fcfbe0392cf888fdba457441ecf5e0475fb63ac5c7\": container with ID starting with c8a3039de6035c530e1e87fcfbe0392cf888fdba457441ecf5e0475fb63ac5c7 not found: ID does not exist" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.628222 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b12a8202-af0b-403e-9767-7cbaeb661cd5-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.735112 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-797ff5c494-4sfjf"] Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.742998 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-797ff5c494-4sfjf"] Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.884413 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b12a8202-af0b-403e-9767-7cbaeb661cd5" path="/var/lib/kubelet/pods/b12a8202-af0b-403e-9767-7cbaeb661cd5/volumes" Feb 14 11:28:01 crc kubenswrapper[4904]: I0214 11:28:01.979220 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.012175 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5c6f7968ff-b5wb6"] Feb 14 11:28:04 crc kubenswrapper[4904]: E0214 11:28:04.012896 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12a8202-af0b-403e-9767-7cbaeb661cd5" containerName="barbican-api-log" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.012910 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12a8202-af0b-403e-9767-7cbaeb661cd5" containerName="barbican-api-log" Feb 14 11:28:04 crc kubenswrapper[4904]: E0214 11:28:04.012935 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12a8202-af0b-403e-9767-7cbaeb661cd5" containerName="barbican-api" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.012941 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12a8202-af0b-403e-9767-7cbaeb661cd5" containerName="barbican-api" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.013128 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12a8202-af0b-403e-9767-7cbaeb661cd5" containerName="barbican-api" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.013153 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12a8202-af0b-403e-9767-7cbaeb661cd5" containerName="barbican-api-log" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.014225 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.016933 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.017153 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.028998 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5c6f7968ff-b5wb6"] Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.031674 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.072426 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b30464d0-5529-441b-8432-0836c20635ea-internal-tls-certs\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.072510 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b30464d0-5529-441b-8432-0836c20635ea-log-httpd\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.072635 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b30464d0-5529-441b-8432-0836c20635ea-run-httpd\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.072681 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhxs7\" (UniqueName: \"kubernetes.io/projected/b30464d0-5529-441b-8432-0836c20635ea-kube-api-access-hhxs7\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.072790 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b30464d0-5529-441b-8432-0836c20635ea-combined-ca-bundle\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.072826 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b30464d0-5529-441b-8432-0836c20635ea-public-tls-certs\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.072987 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b30464d0-5529-441b-8432-0836c20635ea-config-data\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.073083 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b30464d0-5529-441b-8432-0836c20635ea-etc-swift\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.175119 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b30464d0-5529-441b-8432-0836c20635ea-etc-swift\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.175197 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b30464d0-5529-441b-8432-0836c20635ea-internal-tls-certs\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.175234 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b30464d0-5529-441b-8432-0836c20635ea-log-httpd\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.175259 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b30464d0-5529-441b-8432-0836c20635ea-run-httpd\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.175277 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhxs7\" (UniqueName: \"kubernetes.io/projected/b30464d0-5529-441b-8432-0836c20635ea-kube-api-access-hhxs7\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.175316 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b30464d0-5529-441b-8432-0836c20635ea-combined-ca-bundle\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.175338 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b30464d0-5529-441b-8432-0836c20635ea-public-tls-certs\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.175379 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b30464d0-5529-441b-8432-0836c20635ea-config-data\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.175725 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b30464d0-5529-441b-8432-0836c20635ea-run-httpd\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.176433 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b30464d0-5529-441b-8432-0836c20635ea-log-httpd\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.182876 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b30464d0-5529-441b-8432-0836c20635ea-etc-swift\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.183623 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b30464d0-5529-441b-8432-0836c20635ea-internal-tls-certs\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.185009 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b30464d0-5529-441b-8432-0836c20635ea-config-data\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.191204 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhxs7\" (UniqueName: \"kubernetes.io/projected/b30464d0-5529-441b-8432-0836c20635ea-kube-api-access-hhxs7\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.191415 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b30464d0-5529-441b-8432-0836c20635ea-public-tls-certs\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.202529 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b30464d0-5529-441b-8432-0836c20635ea-combined-ca-bundle\") pod \"swift-proxy-5c6f7968ff-b5wb6\" (UID: \"b30464d0-5529-441b-8432-0836c20635ea\") " pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:04 crc kubenswrapper[4904]: I0214 11:28:04.331445 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:06 crc kubenswrapper[4904]: I0214 11:28:06.774659 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:28:06 crc kubenswrapper[4904]: I0214 11:28:06.775117 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="068d8b3e-9bde-4ac2-bcab-1c3e46670bba" containerName="glance-log" containerID="cri-o://3cb6e0644c18f4726141d54faf948ed7ccebf0a4e5ee9165ae1ba3faa19dd21b" gracePeriod=30 Feb 14 11:28:06 crc kubenswrapper[4904]: I0214 11:28:06.775559 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="068d8b3e-9bde-4ac2-bcab-1c3e46670bba" containerName="glance-httpd" containerID="cri-o://1622edf726b9b4d539db9c2f4fa2c5488024e9c391291015fa1f502a76ef8a9b" gracePeriod=30 Feb 14 11:28:07 crc kubenswrapper[4904]: I0214 11:28:07.457546 4904 generic.go:334] "Generic (PLEG): container finished" podID="068d8b3e-9bde-4ac2-bcab-1c3e46670bba" containerID="3cb6e0644c18f4726141d54faf948ed7ccebf0a4e5ee9165ae1ba3faa19dd21b" exitCode=143 Feb 14 11:28:07 crc kubenswrapper[4904]: I0214 11:28:07.457946 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"068d8b3e-9bde-4ac2-bcab-1c3e46670bba","Type":"ContainerDied","Data":"3cb6e0644c18f4726141d54faf948ed7ccebf0a4e5ee9165ae1ba3faa19dd21b"} Feb 14 11:28:07 crc kubenswrapper[4904]: I0214 11:28:07.578562 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 14 11:28:07 crc kubenswrapper[4904]: I0214 11:28:07.741212 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7856f6fd6f-4rhbv" Feb 14 11:28:07 crc kubenswrapper[4904]: I0214 11:28:07.802380 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-55f65b5df6-s2ts6"] Feb 14 11:28:07 crc kubenswrapper[4904]: I0214 11:28:07.802640 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-55f65b5df6-s2ts6" podUID="6db2b9f3-7f93-47e8-b493-eaed893c2b50" containerName="neutron-api" containerID="cri-o://ae3fafc76a52c56cbeeeab7e7d6b63689bc9a8947fc5707d06ef01e25c10fc50" gracePeriod=30 Feb 14 11:28:07 crc kubenswrapper[4904]: I0214 11:28:07.803050 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-55f65b5df6-s2ts6" podUID="6db2b9f3-7f93-47e8-b493-eaed893c2b50" containerName="neutron-httpd" containerID="cri-o://bd9338ff572e4987dc33310a36fef6ebc04cc99afbddefc451c4a431ad71fa42" gracePeriod=30 Feb 14 11:28:08 crc kubenswrapper[4904]: I0214 11:28:08.494952 4904 generic.go:334] "Generic (PLEG): container finished" podID="77818f86-c391-47b6-afff-5bf2831b9328" containerID="dd3d3c3307a079686348e28c4f4826c7b4b8390ad9774ebf70d9ef1a082f06d6" exitCode=137 Feb 14 11:28:08 crc kubenswrapper[4904]: I0214 11:28:08.495315 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"77818f86-c391-47b6-afff-5bf2831b9328","Type":"ContainerDied","Data":"dd3d3c3307a079686348e28c4f4826c7b4b8390ad9774ebf70d9ef1a082f06d6"} Feb 14 11:28:08 crc kubenswrapper[4904]: I0214 11:28:08.506251 4904 generic.go:334] "Generic (PLEG): container finished" podID="6db2b9f3-7f93-47e8-b493-eaed893c2b50" containerID="bd9338ff572e4987dc33310a36fef6ebc04cc99afbddefc451c4a431ad71fa42" exitCode=0 Feb 14 11:28:08 crc kubenswrapper[4904]: I0214 11:28:08.506306 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55f65b5df6-s2ts6" event={"ID":"6db2b9f3-7f93-47e8-b493-eaed893c2b50","Type":"ContainerDied","Data":"bd9338ff572e4987dc33310a36fef6ebc04cc99afbddefc451c4a431ad71fa42"} Feb 14 11:28:08 crc kubenswrapper[4904]: I0214 11:28:08.559424 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:28:08 crc kubenswrapper[4904]: I0214 11:28:08.559732 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2a959004-612f-45ee-82ee-9d1fe7af7eba" containerName="glance-httpd" containerID="cri-o://7233d6630e459d23bd14e59aaaf7361d14eddf1e0c57aa2a7a93459663d61fbd" gracePeriod=30 Feb 14 11:28:08 crc kubenswrapper[4904]: I0214 11:28:08.560105 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2a959004-612f-45ee-82ee-9d1fe7af7eba" containerName="glance-log" containerID="cri-o://60066d7089ba0c88d30e7c41fb469f1d08af664bc554bff90b08522c23dcb662" gracePeriod=30 Feb 14 11:28:08 crc kubenswrapper[4904]: I0214 11:28:08.630563 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5c6f7968ff-b5wb6"] Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.093122 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.185451 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/77818f86-c391-47b6-afff-5bf2831b9328-etc-machine-id\") pod \"77818f86-c391-47b6-afff-5bf2831b9328\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.185797 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-config-data-custom\") pod \"77818f86-c391-47b6-afff-5bf2831b9328\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.185879 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77818f86-c391-47b6-afff-5bf2831b9328-logs\") pod \"77818f86-c391-47b6-afff-5bf2831b9328\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.185904 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drlvn\" (UniqueName: \"kubernetes.io/projected/77818f86-c391-47b6-afff-5bf2831b9328-kube-api-access-drlvn\") pod \"77818f86-c391-47b6-afff-5bf2831b9328\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.185994 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-combined-ca-bundle\") pod \"77818f86-c391-47b6-afff-5bf2831b9328\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.186044 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-scripts\") pod \"77818f86-c391-47b6-afff-5bf2831b9328\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.186098 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-config-data\") pod \"77818f86-c391-47b6-afff-5bf2831b9328\" (UID: \"77818f86-c391-47b6-afff-5bf2831b9328\") " Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.185671 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77818f86-c391-47b6-afff-5bf2831b9328-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "77818f86-c391-47b6-afff-5bf2831b9328" (UID: "77818f86-c391-47b6-afff-5bf2831b9328"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.186769 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77818f86-c391-47b6-afff-5bf2831b9328-logs" (OuterVolumeSpecName: "logs") pod "77818f86-c391-47b6-afff-5bf2831b9328" (UID: "77818f86-c391-47b6-afff-5bf2831b9328"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.195989 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "77818f86-c391-47b6-afff-5bf2831b9328" (UID: "77818f86-c391-47b6-afff-5bf2831b9328"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.202403 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77818f86-c391-47b6-afff-5bf2831b9328-kube-api-access-drlvn" (OuterVolumeSpecName: "kube-api-access-drlvn") pod "77818f86-c391-47b6-afff-5bf2831b9328" (UID: "77818f86-c391-47b6-afff-5bf2831b9328"). InnerVolumeSpecName "kube-api-access-drlvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.207960 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-scripts" (OuterVolumeSpecName: "scripts") pod "77818f86-c391-47b6-afff-5bf2831b9328" (UID: "77818f86-c391-47b6-afff-5bf2831b9328"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.259708 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77818f86-c391-47b6-afff-5bf2831b9328" (UID: "77818f86-c391-47b6-afff-5bf2831b9328"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.288188 4904 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77818f86-c391-47b6-afff-5bf2831b9328-logs\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.288217 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drlvn\" (UniqueName: \"kubernetes.io/projected/77818f86-c391-47b6-afff-5bf2831b9328-kube-api-access-drlvn\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.288227 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.288236 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.288244 4904 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/77818f86-c391-47b6-afff-5bf2831b9328-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.288251 4904 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.373519 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-config-data" (OuterVolumeSpecName: "config-data") pod "77818f86-c391-47b6-afff-5bf2831b9328" (UID: "77818f86-c391-47b6-afff-5bf2831b9328"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.390091 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77818f86-c391-47b6-afff-5bf2831b9328-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.551049 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5c6f7968ff-b5wb6" event={"ID":"b30464d0-5529-441b-8432-0836c20635ea","Type":"ContainerStarted","Data":"3197add7863dd6f5bb4c92966d379da28671a9df8e24ffa1bfb018b640794904"} Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.551111 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5c6f7968ff-b5wb6" event={"ID":"b30464d0-5529-441b-8432-0836c20635ea","Type":"ContainerStarted","Data":"b2bd42d4f29811970604db5e226e0eb19422283d126b128388e1648cefc05022"} Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.552808 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ab9aab3c-a75c-4f0a-9125-623bb22314d0","Type":"ContainerStarted","Data":"883ea9355fe780c493cdcaa1b3ce61592f1a24863c8ee4868e498acc2769be86"} Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.554133 4904 generic.go:334] "Generic (PLEG): container finished" podID="2a959004-612f-45ee-82ee-9d1fe7af7eba" containerID="60066d7089ba0c88d30e7c41fb469f1d08af664bc554bff90b08522c23dcb662" exitCode=143 Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.554182 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2a959004-612f-45ee-82ee-9d1fe7af7eba","Type":"ContainerDied","Data":"60066d7089ba0c88d30e7c41fb469f1d08af664bc554bff90b08522c23dcb662"} Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.559446 4904 generic.go:334] "Generic (PLEG): container finished" podID="cfb252c0-ffb2-4d87-9c4e-b56cc2070d71" containerID="14da3471fa5fd4d6ec42e957717617fb2dea34019a855f12f66753cef241842c" exitCode=137 Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.559497 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bc9d549b8-hdn6n" event={"ID":"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71","Type":"ContainerDied","Data":"14da3471fa5fd4d6ec42e957717617fb2dea34019a855f12f66753cef241842c"} Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.559531 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bc9d549b8-hdn6n" event={"ID":"cfb252c0-ffb2-4d87-9c4e-b56cc2070d71","Type":"ContainerStarted","Data":"f386dd191cfc1be8036723adce1263ce1baf33fbe3f1089ca990ba736b2a5636"} Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.570238 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.160801143 podStartE2EDuration="13.570223162s" podCreationTimestamp="2026-02-14 11:27:56 +0000 UTC" firstStartedPulling="2026-02-14 11:27:57.304986319 +0000 UTC m=+1068.117750980" lastFinishedPulling="2026-02-14 11:28:08.714408338 +0000 UTC m=+1079.527172999" observedRunningTime="2026-02-14 11:28:09.56978575 +0000 UTC m=+1080.382550401" watchObservedRunningTime="2026-02-14 11:28:09.570223162 +0000 UTC m=+1080.382987823" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.579588 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.579654 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"77818f86-c391-47b6-afff-5bf2831b9328","Type":"ContainerDied","Data":"f853eafe1f5b6bee6df9fa4477ebdb1430c7b9d885ce5bcdbcea02ebe3f335d9"} Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.579693 4904 scope.go:117] "RemoveContainer" containerID="dd3d3c3307a079686348e28c4f4826c7b4b8390ad9774ebf70d9ef1a082f06d6" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.617149 4904 generic.go:334] "Generic (PLEG): container finished" podID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerID="609e50be7f47738cb94497fefee796e278d503fdc50869f2e264123ccbb09fa6" exitCode=137 Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.617199 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-687657b496-tz596" event={"ID":"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393","Type":"ContainerDied","Data":"609e50be7f47738cb94497fefee796e278d503fdc50869f2e264123ccbb09fa6"} Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.623293 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.644846 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.669914 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 14 11:28:09 crc kubenswrapper[4904]: E0214 11:28:09.670375 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77818f86-c391-47b6-afff-5bf2831b9328" containerName="cinder-api-log" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.670394 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="77818f86-c391-47b6-afff-5bf2831b9328" containerName="cinder-api-log" Feb 14 11:28:09 crc kubenswrapper[4904]: E0214 11:28:09.670416 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77818f86-c391-47b6-afff-5bf2831b9328" containerName="cinder-api" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.670424 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="77818f86-c391-47b6-afff-5bf2831b9328" containerName="cinder-api" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.670615 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="77818f86-c391-47b6-afff-5bf2831b9328" containerName="cinder-api" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.670636 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="77818f86-c391-47b6-afff-5bf2831b9328" containerName="cinder-api-log" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.671561 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.674110 4904 scope.go:117] "RemoveContainer" containerID="e07211442907acafa7321466c70ea387e33038cb4a3b210324aa5761de4e6ff1" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.674384 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.674386 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.677722 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.707086 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.803573 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dd766f2e-4ddd-474b-862e-e0e568132e9c-config-data-custom\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.803969 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dd766f2e-4ddd-474b-862e-e0e568132e9c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.804025 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd766f2e-4ddd-474b-862e-e0e568132e9c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.804059 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd766f2e-4ddd-474b-862e-e0e568132e9c-config-data\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.804103 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd766f2e-4ddd-474b-862e-e0e568132e9c-logs\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.804151 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd766f2e-4ddd-474b-862e-e0e568132e9c-scripts\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.804179 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd766f2e-4ddd-474b-862e-e0e568132e9c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.804204 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c26dw\" (UniqueName: \"kubernetes.io/projected/dd766f2e-4ddd-474b-862e-e0e568132e9c-kube-api-access-c26dw\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.804235 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd766f2e-4ddd-474b-862e-e0e568132e9c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.861559 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77818f86-c391-47b6-afff-5bf2831b9328" path="/var/lib/kubelet/pods/77818f86-c391-47b6-afff-5bf2831b9328/volumes" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.905930 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd766f2e-4ddd-474b-862e-e0e568132e9c-config-data\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.906025 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd766f2e-4ddd-474b-862e-e0e568132e9c-logs\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.906075 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd766f2e-4ddd-474b-862e-e0e568132e9c-scripts\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.906097 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd766f2e-4ddd-474b-862e-e0e568132e9c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.906115 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c26dw\" (UniqueName: \"kubernetes.io/projected/dd766f2e-4ddd-474b-862e-e0e568132e9c-kube-api-access-c26dw\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.906139 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd766f2e-4ddd-474b-862e-e0e568132e9c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.906164 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dd766f2e-4ddd-474b-862e-e0e568132e9c-config-data-custom\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.906237 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dd766f2e-4ddd-474b-862e-e0e568132e9c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.906268 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd766f2e-4ddd-474b-862e-e0e568132e9c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.907287 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd766f2e-4ddd-474b-862e-e0e568132e9c-logs\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.913716 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dd766f2e-4ddd-474b-862e-e0e568132e9c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.920078 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.920332 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.921610 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.925354 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c26dw\" (UniqueName: \"kubernetes.io/projected/dd766f2e-4ddd-474b-862e-e0e568132e9c-kube-api-access-c26dw\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.926418 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd766f2e-4ddd-474b-862e-e0e568132e9c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.941415 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd766f2e-4ddd-474b-862e-e0e568132e9c-scripts\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.941531 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd766f2e-4ddd-474b-862e-e0e568132e9c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.941649 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dd766f2e-4ddd-474b-862e-e0e568132e9c-config-data-custom\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.956063 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd766f2e-4ddd-474b-862e-e0e568132e9c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.962293 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd766f2e-4ddd-474b-862e-e0e568132e9c-config-data\") pod \"cinder-api-0\" (UID: \"dd766f2e-4ddd-474b-862e-e0e568132e9c\") " pod="openstack/cinder-api-0" Feb 14 11:28:09 crc kubenswrapper[4904]: I0214 11:28:09.999007 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 14 11:28:10 crc kubenswrapper[4904]: I0214 11:28:10.456638 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 14 11:28:10 crc kubenswrapper[4904]: W0214 11:28:10.466418 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd766f2e_4ddd_474b_862e_e0e568132e9c.slice/crio-f75d9c493d2b3fb2ff1f6d3876297e30e8d0e33bc59bd9644dc248f9bfe46230 WatchSource:0}: Error finding container f75d9c493d2b3fb2ff1f6d3876297e30e8d0e33bc59bd9644dc248f9bfe46230: Status 404 returned error can't find the container with id f75d9c493d2b3fb2ff1f6d3876297e30e8d0e33bc59bd9644dc248f9bfe46230 Feb 14 11:28:10 crc kubenswrapper[4904]: I0214 11:28:10.630515 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5c6f7968ff-b5wb6" event={"ID":"b30464d0-5529-441b-8432-0836c20635ea","Type":"ContainerStarted","Data":"70f9a0235008ed093a974d91533c41bbc6e646ac8bb7410e7c2e98f6c6758000"} Feb 14 11:28:10 crc kubenswrapper[4904]: I0214 11:28:10.630876 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:10 crc kubenswrapper[4904]: I0214 11:28:10.633878 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"dd766f2e-4ddd-474b-862e-e0e568132e9c","Type":"ContainerStarted","Data":"f75d9c493d2b3fb2ff1f6d3876297e30e8d0e33bc59bd9644dc248f9bfe46230"} Feb 14 11:28:10 crc kubenswrapper[4904]: I0214 11:28:10.635816 4904 generic.go:334] "Generic (PLEG): container finished" podID="068d8b3e-9bde-4ac2-bcab-1c3e46670bba" containerID="1622edf726b9b4d539db9c2f4fa2c5488024e9c391291015fa1f502a76ef8a9b" exitCode=0 Feb 14 11:28:10 crc kubenswrapper[4904]: I0214 11:28:10.635863 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"068d8b3e-9bde-4ac2-bcab-1c3e46670bba","Type":"ContainerDied","Data":"1622edf726b9b4d539db9c2f4fa2c5488024e9c391291015fa1f502a76ef8a9b"} Feb 14 11:28:10 crc kubenswrapper[4904]: I0214 11:28:10.641650 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="068d8b3e-9bde-4ac2-bcab-1c3e46670bba" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.153:9292/healthcheck\": dial tcp 10.217.0.153:9292: connect: connection refused" Feb 14 11:28:10 crc kubenswrapper[4904]: I0214 11:28:10.641681 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="068d8b3e-9bde-4ac2-bcab-1c3e46670bba" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.153:9292/healthcheck\": dial tcp 10.217.0.153:9292: connect: connection refused" Feb 14 11:28:10 crc kubenswrapper[4904]: I0214 11:28:10.648420 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-687657b496-tz596" event={"ID":"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393","Type":"ContainerStarted","Data":"a6c1837a2344350a20faa0c4c53cf29edd5c3937922bb9a13dd8aa63719d0470"} Feb 14 11:28:10 crc kubenswrapper[4904]: I0214 11:28:10.655205 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5c6f7968ff-b5wb6" podStartSLOduration=7.655189496 podStartE2EDuration="7.655189496s" podCreationTimestamp="2026-02-14 11:28:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:28:10.653942622 +0000 UTC m=+1081.466707303" watchObservedRunningTime="2026-02-14 11:28:10.655189496 +0000 UTC m=+1081.467954157" Feb 14 11:28:11 crc kubenswrapper[4904]: I0214 11:28:11.677007 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"dd766f2e-4ddd-474b-862e-e0e568132e9c","Type":"ContainerStarted","Data":"97d346335a0958eaead7d1442688ba1a7b69d9742363c86fc9eee0076592ea05"} Feb 14 11:28:11 crc kubenswrapper[4904]: I0214 11:28:11.677407 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:11 crc kubenswrapper[4904]: I0214 11:28:11.743020 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="2a959004-612f-45ee-82ee-9d1fe7af7eba" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.154:9292/healthcheck\": read tcp 10.217.0.2:57648->10.217.0.154:9292: read: connection reset by peer" Feb 14 11:28:11 crc kubenswrapper[4904]: I0214 11:28:11.743034 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="2a959004-612f-45ee-82ee-9d1fe7af7eba" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.154:9292/healthcheck\": read tcp 10.217.0.2:57640->10.217.0.154:9292: read: connection reset by peer" Feb 14 11:28:11 crc kubenswrapper[4904]: I0214 11:28:11.925889 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:11 crc kubenswrapper[4904]: I0214 11:28:11.926176 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerName="ceilometer-central-agent" containerID="cri-o://0fc7805dae0f35163979dad6a3dc270a83092837b7fcf764383338be3e0d6038" gracePeriod=30 Feb 14 11:28:11 crc kubenswrapper[4904]: I0214 11:28:11.926656 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerName="proxy-httpd" containerID="cri-o://aff6f3d23a47f62fe7f814538d959f2633736b7949ea773881880c1e95b1d205" gracePeriod=30 Feb 14 11:28:11 crc kubenswrapper[4904]: I0214 11:28:11.926702 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerName="sg-core" containerID="cri-o://e19677930607ed94dd63e028a05e75517c49461ca3589ebee5ec18a9cecac409" gracePeriod=30 Feb 14 11:28:11 crc kubenswrapper[4904]: I0214 11:28:11.926735 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerName="ceilometer-notification-agent" containerID="cri-o://9dbb902d96c46ece6c735688777a0dceda0d33dd9089c12d6ac517b70be69e1f" gracePeriod=30 Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.268055 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.383882 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-scripts\") pod \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.383941 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.384551 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-combined-ca-bundle\") pod \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.384639 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-config-data\") pod \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.384688 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-logs\") pod \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.384849 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-public-tls-certs\") pod \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.384886 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qll2q\" (UniqueName: \"kubernetes.io/projected/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-kube-api-access-qll2q\") pod \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.384991 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-httpd-run\") pod \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.385704 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "068d8b3e-9bde-4ac2-bcab-1c3e46670bba" (UID: "068d8b3e-9bde-4ac2-bcab-1c3e46670bba"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.389779 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-logs" (OuterVolumeSpecName: "logs") pod "068d8b3e-9bde-4ac2-bcab-1c3e46670bba" (UID: "068d8b3e-9bde-4ac2-bcab-1c3e46670bba"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.397588 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-kube-api-access-qll2q" (OuterVolumeSpecName: "kube-api-access-qll2q") pod "068d8b3e-9bde-4ac2-bcab-1c3e46670bba" (UID: "068d8b3e-9bde-4ac2-bcab-1c3e46670bba"). InnerVolumeSpecName "kube-api-access-qll2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.402348 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-scripts" (OuterVolumeSpecName: "scripts") pod "068d8b3e-9bde-4ac2-bcab-1c3e46670bba" (UID: "068d8b3e-9bde-4ac2-bcab-1c3e46670bba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.415385 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "068d8b3e-9bde-4ac2-bcab-1c3e46670bba" (UID: "068d8b3e-9bde-4ac2-bcab-1c3e46670bba"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.487434 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qll2q\" (UniqueName: \"kubernetes.io/projected/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-kube-api-access-qll2q\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.487472 4904 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.487484 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.487512 4904 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.487523 4904 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-logs\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.509993 4904 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.588990 4904 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.686584 4904 generic.go:334] "Generic (PLEG): container finished" podID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerID="e19677930607ed94dd63e028a05e75517c49461ca3589ebee5ec18a9cecac409" exitCode=2 Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.686657 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bccb523-3f10-41ba-aeba-0cd9675d20e2","Type":"ContainerDied","Data":"e19677930607ed94dd63e028a05e75517c49461ca3589ebee5ec18a9cecac409"} Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.689157 4904 generic.go:334] "Generic (PLEG): container finished" podID="2a959004-612f-45ee-82ee-9d1fe7af7eba" containerID="7233d6630e459d23bd14e59aaaf7361d14eddf1e0c57aa2a7a93459663d61fbd" exitCode=0 Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.689235 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2a959004-612f-45ee-82ee-9d1fe7af7eba","Type":"ContainerDied","Data":"7233d6630e459d23bd14e59aaaf7361d14eddf1e0c57aa2a7a93459663d61fbd"} Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.691388 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"068d8b3e-9bde-4ac2-bcab-1c3e46670bba","Type":"ContainerDied","Data":"3eb7ab2eba0e676c34a243469aff7c63fb338818f5ff6e2d94be24861e684ac1"} Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.691413 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.691433 4904 scope.go:117] "RemoveContainer" containerID="1622edf726b9b4d539db9c2f4fa2c5488024e9c391291015fa1f502a76ef8a9b" Feb 14 11:28:12 crc kubenswrapper[4904]: E0214 11:28:12.787699 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-public-tls-certs podName:068d8b3e-9bde-4ac2-bcab-1c3e46670bba nodeName:}" failed. No retries permitted until 2026-02-14 11:28:13.287667455 +0000 UTC m=+1084.100432126 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "public-tls-certs" (UniqueName: "kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-public-tls-certs") pod "068d8b3e-9bde-4ac2-bcab-1c3e46670bba" (UID: "068d8b3e-9bde-4ac2-bcab-1c3e46670bba") : error deleting /var/lib/kubelet/pods/068d8b3e-9bde-4ac2-bcab-1c3e46670bba/volume-subpaths: remove /var/lib/kubelet/pods/068d8b3e-9bde-4ac2-bcab-1c3e46670bba/volume-subpaths: no such file or directory Feb 14 11:28:12 crc kubenswrapper[4904]: E0214 11:28:12.787738 4904 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-config-data podName:068d8b3e-9bde-4ac2-bcab-1c3e46670bba nodeName:}" failed. No retries permitted until 2026-02-14 11:28:13.287731557 +0000 UTC m=+1084.100496208 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-config-data") pod "068d8b3e-9bde-4ac2-bcab-1c3e46670bba" (UID: "068d8b3e-9bde-4ac2-bcab-1c3e46670bba") : error deleting /var/lib/kubelet/pods/068d8b3e-9bde-4ac2-bcab-1c3e46670bba/volume-subpaths: remove /var/lib/kubelet/pods/068d8b3e-9bde-4ac2-bcab-1c3e46670bba/volume-subpaths: no such file or directory Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.791652 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "068d8b3e-9bde-4ac2-bcab-1c3e46670bba" (UID: "068d8b3e-9bde-4ac2-bcab-1c3e46670bba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.792336 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:12 crc kubenswrapper[4904]: I0214 11:28:12.932294 4904 scope.go:117] "RemoveContainer" containerID="3cb6e0644c18f4726141d54faf948ed7ccebf0a4e5ee9165ae1ba3faa19dd21b" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.212548 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.303897 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-config-data\") pod \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.303971 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2a959004-612f-45ee-82ee-9d1fe7af7eba-httpd-run\") pod \"2a959004-612f-45ee-82ee-9d1fe7af7eba\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.304016 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a959004-612f-45ee-82ee-9d1fe7af7eba-logs\") pod \"2a959004-612f-45ee-82ee-9d1fe7af7eba\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.304136 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsjkv\" (UniqueName: \"kubernetes.io/projected/2a959004-612f-45ee-82ee-9d1fe7af7eba-kube-api-access-dsjkv\") pod \"2a959004-612f-45ee-82ee-9d1fe7af7eba\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.304174 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-combined-ca-bundle\") pod \"2a959004-612f-45ee-82ee-9d1fe7af7eba\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.304219 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-public-tls-certs\") pod \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\" (UID: \"068d8b3e-9bde-4ac2-bcab-1c3e46670bba\") " Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.304258 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"2a959004-612f-45ee-82ee-9d1fe7af7eba\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.304337 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-internal-tls-certs\") pod \"2a959004-612f-45ee-82ee-9d1fe7af7eba\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.304355 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-config-data\") pod \"2a959004-612f-45ee-82ee-9d1fe7af7eba\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.304393 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-scripts\") pod \"2a959004-612f-45ee-82ee-9d1fe7af7eba\" (UID: \"2a959004-612f-45ee-82ee-9d1fe7af7eba\") " Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.304475 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a959004-612f-45ee-82ee-9d1fe7af7eba-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2a959004-612f-45ee-82ee-9d1fe7af7eba" (UID: "2a959004-612f-45ee-82ee-9d1fe7af7eba"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.304964 4904 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2a959004-612f-45ee-82ee-9d1fe7af7eba-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.305725 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a959004-612f-45ee-82ee-9d1fe7af7eba-logs" (OuterVolumeSpecName: "logs") pod "2a959004-612f-45ee-82ee-9d1fe7af7eba" (UID: "2a959004-612f-45ee-82ee-9d1fe7af7eba"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.329160 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-config-data" (OuterVolumeSpecName: "config-data") pod "068d8b3e-9bde-4ac2-bcab-1c3e46670bba" (UID: "068d8b3e-9bde-4ac2-bcab-1c3e46670bba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.352261 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "068d8b3e-9bde-4ac2-bcab-1c3e46670bba" (UID: "068d8b3e-9bde-4ac2-bcab-1c3e46670bba"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.355206 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "2a959004-612f-45ee-82ee-9d1fe7af7eba" (UID: "2a959004-612f-45ee-82ee-9d1fe7af7eba"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.355865 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a959004-612f-45ee-82ee-9d1fe7af7eba-kube-api-access-dsjkv" (OuterVolumeSpecName: "kube-api-access-dsjkv") pod "2a959004-612f-45ee-82ee-9d1fe7af7eba" (UID: "2a959004-612f-45ee-82ee-9d1fe7af7eba"). InnerVolumeSpecName "kube-api-access-dsjkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.382086 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-scripts" (OuterVolumeSpecName: "scripts") pod "2a959004-612f-45ee-82ee-9d1fe7af7eba" (UID: "2a959004-612f-45ee-82ee-9d1fe7af7eba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.391207 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a959004-612f-45ee-82ee-9d1fe7af7eba" (UID: "2a959004-612f-45ee-82ee-9d1fe7af7eba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.403057 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2a959004-612f-45ee-82ee-9d1fe7af7eba" (UID: "2a959004-612f-45ee-82ee-9d1fe7af7eba"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.419381 4904 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.419419 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.419433 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.419443 4904 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a959004-612f-45ee-82ee-9d1fe7af7eba-logs\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.419457 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsjkv\" (UniqueName: \"kubernetes.io/projected/2a959004-612f-45ee-82ee-9d1fe7af7eba-kube-api-access-dsjkv\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.419470 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.419480 4904 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/068d8b3e-9bde-4ac2-bcab-1c3e46670bba-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.419515 4904 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.455371 4904 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.500373 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-config-data" (OuterVolumeSpecName: "config-data") pod "2a959004-612f-45ee-82ee-9d1fe7af7eba" (UID: "2a959004-612f-45ee-82ee-9d1fe7af7eba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.520829 4904 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.520875 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a959004-612f-45ee-82ee-9d1fe7af7eba-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.624343 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.632634 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.674807 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:28:13 crc kubenswrapper[4904]: E0214 11:28:13.675168 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a959004-612f-45ee-82ee-9d1fe7af7eba" containerName="glance-log" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.675184 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a959004-612f-45ee-82ee-9d1fe7af7eba" containerName="glance-log" Feb 14 11:28:13 crc kubenswrapper[4904]: E0214 11:28:13.675196 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="068d8b3e-9bde-4ac2-bcab-1c3e46670bba" containerName="glance-log" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.675203 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="068d8b3e-9bde-4ac2-bcab-1c3e46670bba" containerName="glance-log" Feb 14 11:28:13 crc kubenswrapper[4904]: E0214 11:28:13.675223 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a959004-612f-45ee-82ee-9d1fe7af7eba" containerName="glance-httpd" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.675229 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a959004-612f-45ee-82ee-9d1fe7af7eba" containerName="glance-httpd" Feb 14 11:28:13 crc kubenswrapper[4904]: E0214 11:28:13.675241 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="068d8b3e-9bde-4ac2-bcab-1c3e46670bba" containerName="glance-httpd" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.675247 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="068d8b3e-9bde-4ac2-bcab-1c3e46670bba" containerName="glance-httpd" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.675402 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="068d8b3e-9bde-4ac2-bcab-1c3e46670bba" containerName="glance-log" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.675413 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="068d8b3e-9bde-4ac2-bcab-1c3e46670bba" containerName="glance-httpd" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.675426 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a959004-612f-45ee-82ee-9d1fe7af7eba" containerName="glance-httpd" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.675437 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a959004-612f-45ee-82ee-9d1fe7af7eba" containerName="glance-log" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.676470 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.678981 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.679133 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.689604 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.718132 4904 generic.go:334] "Generic (PLEG): container finished" podID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerID="aff6f3d23a47f62fe7f814538d959f2633736b7949ea773881880c1e95b1d205" exitCode=0 Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.718453 4904 generic.go:334] "Generic (PLEG): container finished" podID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerID="0fc7805dae0f35163979dad6a3dc270a83092837b7fcf764383338be3e0d6038" exitCode=0 Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.718155 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bccb523-3f10-41ba-aeba-0cd9675d20e2","Type":"ContainerDied","Data":"aff6f3d23a47f62fe7f814538d959f2633736b7949ea773881880c1e95b1d205"} Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.718543 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bccb523-3f10-41ba-aeba-0cd9675d20e2","Type":"ContainerDied","Data":"0fc7805dae0f35163979dad6a3dc270a83092837b7fcf764383338be3e0d6038"} Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.724275 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2a959004-612f-45ee-82ee-9d1fe7af7eba","Type":"ContainerDied","Data":"9be91377f1b8adcf629d262e8514611f3212a26d790eb0298ad65b7b91c26a95"} Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.724335 4904 scope.go:117] "RemoveContainer" containerID="7233d6630e459d23bd14e59aaaf7361d14eddf1e0c57aa2a7a93459663d61fbd" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.724462 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.730814 4904 generic.go:334] "Generic (PLEG): container finished" podID="6db2b9f3-7f93-47e8-b493-eaed893c2b50" containerID="ae3fafc76a52c56cbeeeab7e7d6b63689bc9a8947fc5707d06ef01e25c10fc50" exitCode=0 Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.731049 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55f65b5df6-s2ts6" event={"ID":"6db2b9f3-7f93-47e8-b493-eaed893c2b50","Type":"ContainerDied","Data":"ae3fafc76a52c56cbeeeab7e7d6b63689bc9a8947fc5707d06ef01e25c10fc50"} Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.772641 4904 scope.go:117] "RemoveContainer" containerID="60066d7089ba0c88d30e7c41fb469f1d08af664bc554bff90b08522c23dcb662" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.773912 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.783704 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.794581 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.797060 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.801826 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.802617 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.823060 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.827968 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dba92113-8ae6-46ee-b6ca-32da151c96b0-logs\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.828059 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ml5r\" (UniqueName: \"kubernetes.io/projected/dba92113-8ae6-46ee-b6ca-32da151c96b0-kube-api-access-8ml5r\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.828116 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.828176 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dba92113-8ae6-46ee-b6ca-32da151c96b0-scripts\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.828204 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dba92113-8ae6-46ee-b6ca-32da151c96b0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.828278 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba92113-8ae6-46ee-b6ca-32da151c96b0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.828311 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dba92113-8ae6-46ee-b6ca-32da151c96b0-config-data\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.828534 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dba92113-8ae6-46ee-b6ca-32da151c96b0-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.861622 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="068d8b3e-9bde-4ac2-bcab-1c3e46670bba" path="/var/lib/kubelet/pods/068d8b3e-9bde-4ac2-bcab-1c3e46670bba/volumes" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.866144 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a959004-612f-45ee-82ee-9d1fe7af7eba" path="/var/lib/kubelet/pods/2a959004-612f-45ee-82ee-9d1fe7af7eba/volumes" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.930325 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c29c32-71ef-48c9-8e26-428071c13df2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.930387 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87c29c32-71ef-48c9-8e26-428071c13df2-logs\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.930405 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh5hd\" (UniqueName: \"kubernetes.io/projected/87c29c32-71ef-48c9-8e26-428071c13df2-kube-api-access-rh5hd\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.930421 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.930449 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dba92113-8ae6-46ee-b6ca-32da151c96b0-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.930479 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dba92113-8ae6-46ee-b6ca-32da151c96b0-logs\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.930506 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ml5r\" (UniqueName: \"kubernetes.io/projected/dba92113-8ae6-46ee-b6ca-32da151c96b0-kube-api-access-8ml5r\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.930531 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.930544 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dba92113-8ae6-46ee-b6ca-32da151c96b0-scripts\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.930557 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dba92113-8ae6-46ee-b6ca-32da151c96b0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.930576 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c29c32-71ef-48c9-8e26-428071c13df2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.930630 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba92113-8ae6-46ee-b6ca-32da151c96b0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.930884 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dba92113-8ae6-46ee-b6ca-32da151c96b0-config-data\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.930939 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c29c32-71ef-48c9-8e26-428071c13df2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.931023 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/87c29c32-71ef-48c9-8e26-428071c13df2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.931070 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87c29c32-71ef-48c9-8e26-428071c13df2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.931601 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dba92113-8ae6-46ee-b6ca-32da151c96b0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.932409 4904 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.932958 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dba92113-8ae6-46ee-b6ca-32da151c96b0-logs\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.943332 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dba92113-8ae6-46ee-b6ca-32da151c96b0-config-data\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.946487 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dba92113-8ae6-46ee-b6ca-32da151c96b0-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.949551 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba92113-8ae6-46ee-b6ca-32da151c96b0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.950179 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dba92113-8ae6-46ee-b6ca-32da151c96b0-scripts\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.964670 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ml5r\" (UniqueName: \"kubernetes.io/projected/dba92113-8ae6-46ee-b6ca-32da151c96b0-kube-api-access-8ml5r\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:13 crc kubenswrapper[4904]: I0214 11:28:13.999216 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"dba92113-8ae6-46ee-b6ca-32da151c96b0\") " pod="openstack/glance-default-external-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.031248 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.032698 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c29c32-71ef-48c9-8e26-428071c13df2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.032781 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c29c32-71ef-48c9-8e26-428071c13df2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.032858 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/87c29c32-71ef-48c9-8e26-428071c13df2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.032912 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87c29c32-71ef-48c9-8e26-428071c13df2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.032948 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c29c32-71ef-48c9-8e26-428071c13df2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.032994 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87c29c32-71ef-48c9-8e26-428071c13df2-logs\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.033019 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh5hd\" (UniqueName: \"kubernetes.io/projected/87c29c32-71ef-48c9-8e26-428071c13df2-kube-api-access-rh5hd\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.033041 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.033216 4904 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.035790 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87c29c32-71ef-48c9-8e26-428071c13df2-logs\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.036316 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87c29c32-71ef-48c9-8e26-428071c13df2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.036552 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/87c29c32-71ef-48c9-8e26-428071c13df2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.040899 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c29c32-71ef-48c9-8e26-428071c13df2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.042493 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c29c32-71ef-48c9-8e26-428071c13df2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.049473 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c29c32-71ef-48c9-8e26-428071c13df2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.056795 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh5hd\" (UniqueName: \"kubernetes.io/projected/87c29c32-71ef-48c9-8e26-428071c13df2-kube-api-access-rh5hd\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.090771 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"87c29c32-71ef-48c9-8e26-428071c13df2\") " pod="openstack/glance-default-internal-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.117206 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.368890 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-5c6f7968ff-b5wb6" podUID="b30464d0-5529-441b-8432-0836c20635ea" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 503" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.369524 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-5c6f7968ff-b5wb6" podUID="b30464d0-5529-441b-8432-0836c20635ea" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.500344 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.541387 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-combined-ca-bundle\") pod \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.541433 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkr6t\" (UniqueName: \"kubernetes.io/projected/6db2b9f3-7f93-47e8-b493-eaed893c2b50-kube-api-access-nkr6t\") pod \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.541460 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-ovndb-tls-certs\") pod \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.541528 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-httpd-config\") pod \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.541638 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-config\") pod \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\" (UID: \"6db2b9f3-7f93-47e8-b493-eaed893c2b50\") " Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.561116 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6db2b9f3-7f93-47e8-b493-eaed893c2b50-kube-api-access-nkr6t" (OuterVolumeSpecName: "kube-api-access-nkr6t") pod "6db2b9f3-7f93-47e8-b493-eaed893c2b50" (UID: "6db2b9f3-7f93-47e8-b493-eaed893c2b50"). InnerVolumeSpecName "kube-api-access-nkr6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.577588 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "6db2b9f3-7f93-47e8-b493-eaed893c2b50" (UID: "6db2b9f3-7f93-47e8-b493-eaed893c2b50"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.648018 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkr6t\" (UniqueName: \"kubernetes.io/projected/6db2b9f3-7f93-47e8-b493-eaed893c2b50-kube-api-access-nkr6t\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.648049 4904 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.685594 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6db2b9f3-7f93-47e8-b493-eaed893c2b50" (UID: "6db2b9f3-7f93-47e8-b493-eaed893c2b50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.698043 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "6db2b9f3-7f93-47e8-b493-eaed893c2b50" (UID: "6db2b9f3-7f93-47e8-b493-eaed893c2b50"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.698171 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-config" (OuterVolumeSpecName: "config") pod "6db2b9f3-7f93-47e8-b493-eaed893c2b50" (UID: "6db2b9f3-7f93-47e8-b493-eaed893c2b50"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.746817 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55f65b5df6-s2ts6" event={"ID":"6db2b9f3-7f93-47e8-b493-eaed893c2b50","Type":"ContainerDied","Data":"ff0ca90250e14827d90c10778ad1233db548d045ac850cdc51c3071b180d1f67"} Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.746889 4904 scope.go:117] "RemoveContainer" containerID="bd9338ff572e4987dc33310a36fef6ebc04cc99afbddefc451c4a431ad71fa42" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.747002 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55f65b5df6-s2ts6" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.759304 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.759351 4904 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.759362 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6db2b9f3-7f93-47e8-b493-eaed893c2b50-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.762395 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"dd766f2e-4ddd-474b-862e-e0e568132e9c","Type":"ContainerStarted","Data":"13b058479e299df43ee4aa2d085388aa705f53652b8c7b06f743b0b2c7821585"} Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.763364 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.783078 4904 scope.go:117] "RemoveContainer" containerID="ae3fafc76a52c56cbeeeab7e7d6b63689bc9a8947fc5707d06ef01e25c10fc50" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.788720 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-55f65b5df6-s2ts6"] Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.812027 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-55f65b5df6-s2ts6"] Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.813780 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.813762019 podStartE2EDuration="5.813762019s" podCreationTimestamp="2026-02-14 11:28:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:28:14.801159332 +0000 UTC m=+1085.613923993" watchObservedRunningTime="2026-02-14 11:28:14.813762019 +0000 UTC m=+1085.626526670" Feb 14 11:28:14 crc kubenswrapper[4904]: I0214 11:28:14.858269 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.737399 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.782582 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bccb523-3f10-41ba-aeba-0cd9675d20e2-log-httpd\") pod \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.782731 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-sg-core-conf-yaml\") pod \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.782795 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bccb523-3f10-41ba-aeba-0cd9675d20e2-run-httpd\") pod \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.782823 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-config-data\") pod \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.782863 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-scripts\") pod \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.782955 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-combined-ca-bundle\") pod \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.783002 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27qbv\" (UniqueName: \"kubernetes.io/projected/3bccb523-3f10-41ba-aeba-0cd9675d20e2-kube-api-access-27qbv\") pod \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.785311 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bccb523-3f10-41ba-aeba-0cd9675d20e2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3bccb523-3f10-41ba-aeba-0cd9675d20e2" (UID: "3bccb523-3f10-41ba-aeba-0cd9675d20e2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.785597 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bccb523-3f10-41ba-aeba-0cd9675d20e2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3bccb523-3f10-41ba-aeba-0cd9675d20e2" (UID: "3bccb523-3f10-41ba-aeba-0cd9675d20e2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.795010 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-scripts" (OuterVolumeSpecName: "scripts") pod "3bccb523-3f10-41ba-aeba-0cd9675d20e2" (UID: "3bccb523-3f10-41ba-aeba-0cd9675d20e2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.814207 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bccb523-3f10-41ba-aeba-0cd9675d20e2-kube-api-access-27qbv" (OuterVolumeSpecName: "kube-api-access-27qbv") pod "3bccb523-3f10-41ba-aeba-0cd9675d20e2" (UID: "3bccb523-3f10-41ba-aeba-0cd9675d20e2"). InnerVolumeSpecName "kube-api-access-27qbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.844605 4904 generic.go:334] "Generic (PLEG): container finished" podID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerID="9dbb902d96c46ece6c735688777a0dceda0d33dd9089c12d6ac517b70be69e1f" exitCode=0 Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.844741 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.859299 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3bccb523-3f10-41ba-aeba-0cd9675d20e2" (UID: "3bccb523-3f10-41ba-aeba-0cd9675d20e2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.885292 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27qbv\" (UniqueName: \"kubernetes.io/projected/3bccb523-3f10-41ba-aeba-0cd9675d20e2-kube-api-access-27qbv\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.885320 4904 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bccb523-3f10-41ba-aeba-0cd9675d20e2-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.885329 4904 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.885337 4904 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bccb523-3f10-41ba-aeba-0cd9675d20e2-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.885344 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.886360 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6db2b9f3-7f93-47e8-b493-eaed893c2b50" path="/var/lib/kubelet/pods/6db2b9f3-7f93-47e8-b493-eaed893c2b50/volumes" Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.897243 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bccb523-3f10-41ba-aeba-0cd9675d20e2","Type":"ContainerDied","Data":"9dbb902d96c46ece6c735688777a0dceda0d33dd9089c12d6ac517b70be69e1f"} Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.897283 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bccb523-3f10-41ba-aeba-0cd9675d20e2","Type":"ContainerDied","Data":"a075ad508cae7d2e6bfb30368cf5d1ecfe6c32daa8e99efd51101a629719cb6b"} Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.897297 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.897312 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"87c29c32-71ef-48c9-8e26-428071c13df2","Type":"ContainerStarted","Data":"1476d1d870b802ee40dbd15a297cbf51631ac57c40382a7f22397c67a6f4259f"} Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.897329 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"87c29c32-71ef-48c9-8e26-428071c13df2","Type":"ContainerStarted","Data":"a6b0585008628b60e3dcd99f0ef69728b25a56482723aa820649bd67b2648c42"} Feb 14 11:28:15 crc kubenswrapper[4904]: I0214 11:28:15.897339 4904 scope.go:117] "RemoveContainer" containerID="aff6f3d23a47f62fe7f814538d959f2633736b7949ea773881880c1e95b1d205" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.088024 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-config-data" (OuterVolumeSpecName: "config-data") pod "3bccb523-3f10-41ba-aeba-0cd9675d20e2" (UID: "3bccb523-3f10-41ba-aeba-0cd9675d20e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.089638 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-config-data\") pod \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\" (UID: \"3bccb523-3f10-41ba-aeba-0cd9675d20e2\") " Feb 14 11:28:16 crc kubenswrapper[4904]: W0214 11:28:16.090232 4904 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/3bccb523-3f10-41ba-aeba-0cd9675d20e2/volumes/kubernetes.io~secret/config-data Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.090250 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-config-data" (OuterVolumeSpecName: "config-data") pod "3bccb523-3f10-41ba-aeba-0cd9675d20e2" (UID: "3bccb523-3f10-41ba-aeba-0cd9675d20e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.103407 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.133042 4904 scope.go:117] "RemoveContainer" containerID="e19677930607ed94dd63e028a05e75517c49461ca3589ebee5ec18a9cecac409" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.208729 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3bccb523-3f10-41ba-aeba-0cd9675d20e2" (UID: "3bccb523-3f10-41ba-aeba-0cd9675d20e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.308744 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bccb523-3f10-41ba-aeba-0cd9675d20e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.333146 4904 scope.go:117] "RemoveContainer" containerID="9dbb902d96c46ece6c735688777a0dceda0d33dd9089c12d6ac517b70be69e1f" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.337167 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-5c6f7968ff-b5wb6" podUID="b30464d0-5529-441b-8432-0836c20635ea" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.349380 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-5c6f7968ff-b5wb6" podUID="b30464d0-5529-441b-8432-0836c20635ea" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 503" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.409310 4904 scope.go:117] "RemoveContainer" containerID="0fc7805dae0f35163979dad6a3dc270a83092837b7fcf764383338be3e0d6038" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.490223 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.521000 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.539015 4904 scope.go:117] "RemoveContainer" containerID="aff6f3d23a47f62fe7f814538d959f2633736b7949ea773881880c1e95b1d205" Feb 14 11:28:16 crc kubenswrapper[4904]: E0214 11:28:16.539620 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aff6f3d23a47f62fe7f814538d959f2633736b7949ea773881880c1e95b1d205\": container with ID starting with aff6f3d23a47f62fe7f814538d959f2633736b7949ea773881880c1e95b1d205 not found: ID does not exist" containerID="aff6f3d23a47f62fe7f814538d959f2633736b7949ea773881880c1e95b1d205" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.539654 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aff6f3d23a47f62fe7f814538d959f2633736b7949ea773881880c1e95b1d205"} err="failed to get container status \"aff6f3d23a47f62fe7f814538d959f2633736b7949ea773881880c1e95b1d205\": rpc error: code = NotFound desc = could not find container \"aff6f3d23a47f62fe7f814538d959f2633736b7949ea773881880c1e95b1d205\": container with ID starting with aff6f3d23a47f62fe7f814538d959f2633736b7949ea773881880c1e95b1d205 not found: ID does not exist" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.539675 4904 scope.go:117] "RemoveContainer" containerID="e19677930607ed94dd63e028a05e75517c49461ca3589ebee5ec18a9cecac409" Feb 14 11:28:16 crc kubenswrapper[4904]: E0214 11:28:16.540151 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e19677930607ed94dd63e028a05e75517c49461ca3589ebee5ec18a9cecac409\": container with ID starting with e19677930607ed94dd63e028a05e75517c49461ca3589ebee5ec18a9cecac409 not found: ID does not exist" containerID="e19677930607ed94dd63e028a05e75517c49461ca3589ebee5ec18a9cecac409" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.540171 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e19677930607ed94dd63e028a05e75517c49461ca3589ebee5ec18a9cecac409"} err="failed to get container status \"e19677930607ed94dd63e028a05e75517c49461ca3589ebee5ec18a9cecac409\": rpc error: code = NotFound desc = could not find container \"e19677930607ed94dd63e028a05e75517c49461ca3589ebee5ec18a9cecac409\": container with ID starting with e19677930607ed94dd63e028a05e75517c49461ca3589ebee5ec18a9cecac409 not found: ID does not exist" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.540185 4904 scope.go:117] "RemoveContainer" containerID="9dbb902d96c46ece6c735688777a0dceda0d33dd9089c12d6ac517b70be69e1f" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.548897 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:16 crc kubenswrapper[4904]: E0214 11:28:16.549345 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6db2b9f3-7f93-47e8-b493-eaed893c2b50" containerName="neutron-httpd" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.549364 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="6db2b9f3-7f93-47e8-b493-eaed893c2b50" containerName="neutron-httpd" Feb 14 11:28:16 crc kubenswrapper[4904]: E0214 11:28:16.549381 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerName="proxy-httpd" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.549387 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerName="proxy-httpd" Feb 14 11:28:16 crc kubenswrapper[4904]: E0214 11:28:16.549408 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerName="ceilometer-central-agent" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.549415 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerName="ceilometer-central-agent" Feb 14 11:28:16 crc kubenswrapper[4904]: E0214 11:28:16.549428 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerName="ceilometer-notification-agent" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.549434 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerName="ceilometer-notification-agent" Feb 14 11:28:16 crc kubenswrapper[4904]: E0214 11:28:16.549447 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerName="sg-core" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.549452 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerName="sg-core" Feb 14 11:28:16 crc kubenswrapper[4904]: E0214 11:28:16.549463 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6db2b9f3-7f93-47e8-b493-eaed893c2b50" containerName="neutron-api" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.549469 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="6db2b9f3-7f93-47e8-b493-eaed893c2b50" containerName="neutron-api" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.549618 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerName="ceilometer-notification-agent" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.549629 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="6db2b9f3-7f93-47e8-b493-eaed893c2b50" containerName="neutron-api" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.549637 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerName="ceilometer-central-agent" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.549643 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerName="proxy-httpd" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.549654 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="6db2b9f3-7f93-47e8-b493-eaed893c2b50" containerName="neutron-httpd" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.549665 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" containerName="sg-core" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.551290 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: E0214 11:28:16.556243 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dbb902d96c46ece6c735688777a0dceda0d33dd9089c12d6ac517b70be69e1f\": container with ID starting with 9dbb902d96c46ece6c735688777a0dceda0d33dd9089c12d6ac517b70be69e1f not found: ID does not exist" containerID="9dbb902d96c46ece6c735688777a0dceda0d33dd9089c12d6ac517b70be69e1f" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.556279 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dbb902d96c46ece6c735688777a0dceda0d33dd9089c12d6ac517b70be69e1f"} err="failed to get container status \"9dbb902d96c46ece6c735688777a0dceda0d33dd9089c12d6ac517b70be69e1f\": rpc error: code = NotFound desc = could not find container \"9dbb902d96c46ece6c735688777a0dceda0d33dd9089c12d6ac517b70be69e1f\": container with ID starting with 9dbb902d96c46ece6c735688777a0dceda0d33dd9089c12d6ac517b70be69e1f not found: ID does not exist" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.556307 4904 scope.go:117] "RemoveContainer" containerID="0fc7805dae0f35163979dad6a3dc270a83092837b7fcf764383338be3e0d6038" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.560930 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:16 crc kubenswrapper[4904]: E0214 11:28:16.561092 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fc7805dae0f35163979dad6a3dc270a83092837b7fcf764383338be3e0d6038\": container with ID starting with 0fc7805dae0f35163979dad6a3dc270a83092837b7fcf764383338be3e0d6038 not found: ID does not exist" containerID="0fc7805dae0f35163979dad6a3dc270a83092837b7fcf764383338be3e0d6038" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.561117 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fc7805dae0f35163979dad6a3dc270a83092837b7fcf764383338be3e0d6038"} err="failed to get container status \"0fc7805dae0f35163979dad6a3dc270a83092837b7fcf764383338be3e0d6038\": rpc error: code = NotFound desc = could not find container \"0fc7805dae0f35163979dad6a3dc270a83092837b7fcf764383338be3e0d6038\": container with ID starting with 0fc7805dae0f35163979dad6a3dc270a83092837b7fcf764383338be3e0d6038 not found: ID does not exist" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.561284 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.561537 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.623915 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-config-data\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.623963 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.624001 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/562e6e60-d241-4e05-98da-80b3c449919e-log-httpd\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.624026 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k6xq\" (UniqueName: \"kubernetes.io/projected/562e6e60-d241-4e05-98da-80b3c449919e-kube-api-access-5k6xq\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.624082 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.624101 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-scripts\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.624115 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/562e6e60-d241-4e05-98da-80b3c449919e-run-httpd\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.725962 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-config-data\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.726005 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.726043 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/562e6e60-d241-4e05-98da-80b3c449919e-log-httpd\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.726068 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k6xq\" (UniqueName: \"kubernetes.io/projected/562e6e60-d241-4e05-98da-80b3c449919e-kube-api-access-5k6xq\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.726122 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.726139 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-scripts\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.726156 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/562e6e60-d241-4e05-98da-80b3c449919e-run-httpd\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.726560 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/562e6e60-d241-4e05-98da-80b3c449919e-run-httpd\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.727632 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/562e6e60-d241-4e05-98da-80b3c449919e-log-httpd\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.733648 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-scripts\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.733753 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.734438 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.737457 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-config-data\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.746259 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k6xq\" (UniqueName: \"kubernetes.io/projected/562e6e60-d241-4e05-98da-80b3c449919e-kube-api-access-5k6xq\") pod \"ceilometer-0\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.881158 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"87c29c32-71ef-48c9-8e26-428071c13df2","Type":"ContainerStarted","Data":"9d1dcb0017bb2af4f4c6f32c2e939b50abee8a7e9f3e09875655827f439c65ca"} Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.882367 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dba92113-8ae6-46ee-b6ca-32da151c96b0","Type":"ContainerStarted","Data":"8ea3a507b592f4aa61e5d241bde45c36d5c2be4cb1b9407cd195dec26b410294"} Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.897032 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:28:16 crc kubenswrapper[4904]: I0214 11:28:16.905788 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.905772807 podStartE2EDuration="3.905772807s" podCreationTimestamp="2026-02-14 11:28:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:28:16.905270592 +0000 UTC m=+1087.718035253" watchObservedRunningTime="2026-02-14 11:28:16.905772807 +0000 UTC m=+1087.718537468" Feb 14 11:28:17 crc kubenswrapper[4904]: I0214 11:28:17.440496 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:17 crc kubenswrapper[4904]: I0214 11:28:17.849026 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bccb523-3f10-41ba-aeba-0cd9675d20e2" path="/var/lib/kubelet/pods/3bccb523-3f10-41ba-aeba-0cd9675d20e2/volumes" Feb 14 11:28:17 crc kubenswrapper[4904]: I0214 11:28:17.896047 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dba92113-8ae6-46ee-b6ca-32da151c96b0","Type":"ContainerStarted","Data":"84ef129da638b99f7db6992bdde554ebcd8c54637cf86e1f49829c9bd7439be7"} Feb 14 11:28:17 crc kubenswrapper[4904]: I0214 11:28:17.896287 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dba92113-8ae6-46ee-b6ca-32da151c96b0","Type":"ContainerStarted","Data":"f214da4407b8e528edf502961e5b84287aaad201bcecd6fe86b9b50531d31443"} Feb 14 11:28:17 crc kubenswrapper[4904]: I0214 11:28:17.898075 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"562e6e60-d241-4e05-98da-80b3c449919e","Type":"ContainerStarted","Data":"9fe360eee0479af31170fbfb61585e3cdc19ab77ac99961d6e707232b0539953"} Feb 14 11:28:17 crc kubenswrapper[4904]: I0214 11:28:17.929644 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.929594319 podStartE2EDuration="4.929594319s" podCreationTimestamp="2026-02-14 11:28:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:28:17.921142067 +0000 UTC m=+1088.733906728" watchObservedRunningTime="2026-02-14 11:28:17.929594319 +0000 UTC m=+1088.742358980" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.589828 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-msvcb"] Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.591105 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-msvcb" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.608265 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-msvcb"] Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.653861 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-sjsmr"] Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.654827 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-sjsmr" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.670017 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-sjsmr"] Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.671583 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60971f9e-6866-4f58-9686-0eb77268650a-operator-scripts\") pod \"nova-cell0-db-create-sjsmr\" (UID: \"60971f9e-6866-4f58-9686-0eb77268650a\") " pod="openstack/nova-cell0-db-create-sjsmr" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.671760 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d63d7a3-5e12-428e-a9dc-705b28836ca0-operator-scripts\") pod \"nova-api-db-create-msvcb\" (UID: \"1d63d7a3-5e12-428e-a9dc-705b28836ca0\") " pod="openstack/nova-api-db-create-msvcb" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.671896 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwppd\" (UniqueName: \"kubernetes.io/projected/1d63d7a3-5e12-428e-a9dc-705b28836ca0-kube-api-access-mwppd\") pod \"nova-api-db-create-msvcb\" (UID: \"1d63d7a3-5e12-428e-a9dc-705b28836ca0\") " pod="openstack/nova-api-db-create-msvcb" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.672004 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rngr5\" (UniqueName: \"kubernetes.io/projected/60971f9e-6866-4f58-9686-0eb77268650a-kube-api-access-rngr5\") pod \"nova-cell0-db-create-sjsmr\" (UID: \"60971f9e-6866-4f58-9686-0eb77268650a\") " pod="openstack/nova-cell0-db-create-sjsmr" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.754668 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-65a0-account-create-update-lz527"] Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.755722 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-65a0-account-create-update-lz527" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.761762 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.765133 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-65a0-account-create-update-lz527"] Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.773234 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d63d7a3-5e12-428e-a9dc-705b28836ca0-operator-scripts\") pod \"nova-api-db-create-msvcb\" (UID: \"1d63d7a3-5e12-428e-a9dc-705b28836ca0\") " pod="openstack/nova-api-db-create-msvcb" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.773284 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/193a644c-698e-4ad6-a36d-8c19c8ef0d0a-operator-scripts\") pod \"nova-api-65a0-account-create-update-lz527\" (UID: \"193a644c-698e-4ad6-a36d-8c19c8ef0d0a\") " pod="openstack/nova-api-65a0-account-create-update-lz527" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.773335 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwppd\" (UniqueName: \"kubernetes.io/projected/1d63d7a3-5e12-428e-a9dc-705b28836ca0-kube-api-access-mwppd\") pod \"nova-api-db-create-msvcb\" (UID: \"1d63d7a3-5e12-428e-a9dc-705b28836ca0\") " pod="openstack/nova-api-db-create-msvcb" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.773373 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rngr5\" (UniqueName: \"kubernetes.io/projected/60971f9e-6866-4f58-9686-0eb77268650a-kube-api-access-rngr5\") pod \"nova-cell0-db-create-sjsmr\" (UID: \"60971f9e-6866-4f58-9686-0eb77268650a\") " pod="openstack/nova-cell0-db-create-sjsmr" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.773416 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60971f9e-6866-4f58-9686-0eb77268650a-operator-scripts\") pod \"nova-cell0-db-create-sjsmr\" (UID: \"60971f9e-6866-4f58-9686-0eb77268650a\") " pod="openstack/nova-cell0-db-create-sjsmr" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.773454 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4cmz\" (UniqueName: \"kubernetes.io/projected/193a644c-698e-4ad6-a36d-8c19c8ef0d0a-kube-api-access-m4cmz\") pod \"nova-api-65a0-account-create-update-lz527\" (UID: \"193a644c-698e-4ad6-a36d-8c19c8ef0d0a\") " pod="openstack/nova-api-65a0-account-create-update-lz527" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.774255 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d63d7a3-5e12-428e-a9dc-705b28836ca0-operator-scripts\") pod \"nova-api-db-create-msvcb\" (UID: \"1d63d7a3-5e12-428e-a9dc-705b28836ca0\") " pod="openstack/nova-api-db-create-msvcb" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.774502 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60971f9e-6866-4f58-9686-0eb77268650a-operator-scripts\") pod \"nova-cell0-db-create-sjsmr\" (UID: \"60971f9e-6866-4f58-9686-0eb77268650a\") " pod="openstack/nova-cell0-db-create-sjsmr" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.781417 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.782169 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.797363 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwppd\" (UniqueName: \"kubernetes.io/projected/1d63d7a3-5e12-428e-a9dc-705b28836ca0-kube-api-access-mwppd\") pod \"nova-api-db-create-msvcb\" (UID: \"1d63d7a3-5e12-428e-a9dc-705b28836ca0\") " pod="openstack/nova-api-db-create-msvcb" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.804460 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rngr5\" (UniqueName: \"kubernetes.io/projected/60971f9e-6866-4f58-9686-0eb77268650a-kube-api-access-rngr5\") pod \"nova-cell0-db-create-sjsmr\" (UID: \"60971f9e-6866-4f58-9686-0eb77268650a\") " pod="openstack/nova-cell0-db-create-sjsmr" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.868019 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-94bx7"] Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.883926 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-94bx7" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.899086 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4cmz\" (UniqueName: \"kubernetes.io/projected/193a644c-698e-4ad6-a36d-8c19c8ef0d0a-kube-api-access-m4cmz\") pod \"nova-api-65a0-account-create-update-lz527\" (UID: \"193a644c-698e-4ad6-a36d-8c19c8ef0d0a\") " pod="openstack/nova-api-65a0-account-create-update-lz527" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.899268 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/193a644c-698e-4ad6-a36d-8c19c8ef0d0a-operator-scripts\") pod \"nova-api-65a0-account-create-update-lz527\" (UID: \"193a644c-698e-4ad6-a36d-8c19c8ef0d0a\") " pod="openstack/nova-api-65a0-account-create-update-lz527" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.911657 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-msvcb" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.913705 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/193a644c-698e-4ad6-a36d-8c19c8ef0d0a-operator-scripts\") pod \"nova-api-65a0-account-create-update-lz527\" (UID: \"193a644c-698e-4ad6-a36d-8c19c8ef0d0a\") " pod="openstack/nova-api-65a0-account-create-update-lz527" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.919152 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-94bx7"] Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.938631 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"562e6e60-d241-4e05-98da-80b3c449919e","Type":"ContainerStarted","Data":"65653abe18fc0a6a7f40e2f85fafd80b5c9ac25b277b21ddc74f8fab2e346a63"} Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.949707 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-687657b496-tz596" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.949954 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-687657b496-tz596" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.971385 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-sjsmr" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.971772 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-3e19-account-create-update-bxrv4"] Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.973121 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4cmz\" (UniqueName: \"kubernetes.io/projected/193a644c-698e-4ad6-a36d-8c19c8ef0d0a-kube-api-access-m4cmz\") pod \"nova-api-65a0-account-create-update-lz527\" (UID: \"193a644c-698e-4ad6-a36d-8c19c8ef0d0a\") " pod="openstack/nova-api-65a0-account-create-update-lz527" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.975218 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3e19-account-create-update-bxrv4" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.978452 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 14 11:28:18 crc kubenswrapper[4904]: I0214 11:28:18.981933 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-3e19-account-create-update-bxrv4"] Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.001486 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnvqb\" (UniqueName: \"kubernetes.io/projected/e8121d8f-c473-4818-80f2-1335e44a7955-kube-api-access-vnvqb\") pod \"nova-cell0-3e19-account-create-update-bxrv4\" (UID: \"e8121d8f-c473-4818-80f2-1335e44a7955\") " pod="openstack/nova-cell0-3e19-account-create-update-bxrv4" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.001578 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6ggj\" (UniqueName: \"kubernetes.io/projected/43b09009-a1b8-4f5f-8875-8a01b9672adb-kube-api-access-s6ggj\") pod \"nova-cell1-db-create-94bx7\" (UID: \"43b09009-a1b8-4f5f-8875-8a01b9672adb\") " pod="openstack/nova-cell1-db-create-94bx7" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.001677 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43b09009-a1b8-4f5f-8875-8a01b9672adb-operator-scripts\") pod \"nova-cell1-db-create-94bx7\" (UID: \"43b09009-a1b8-4f5f-8875-8a01b9672adb\") " pod="openstack/nova-cell1-db-create-94bx7" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.001753 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8121d8f-c473-4818-80f2-1335e44a7955-operator-scripts\") pod \"nova-cell0-3e19-account-create-update-bxrv4\" (UID: \"e8121d8f-c473-4818-80f2-1335e44a7955\") " pod="openstack/nova-cell0-3e19-account-create-update-bxrv4" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.072616 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-65a0-account-create-update-lz527" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.104184 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnvqb\" (UniqueName: \"kubernetes.io/projected/e8121d8f-c473-4818-80f2-1335e44a7955-kube-api-access-vnvqb\") pod \"nova-cell0-3e19-account-create-update-bxrv4\" (UID: \"e8121d8f-c473-4818-80f2-1335e44a7955\") " pod="openstack/nova-cell0-3e19-account-create-update-bxrv4" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.104250 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6ggj\" (UniqueName: \"kubernetes.io/projected/43b09009-a1b8-4f5f-8875-8a01b9672adb-kube-api-access-s6ggj\") pod \"nova-cell1-db-create-94bx7\" (UID: \"43b09009-a1b8-4f5f-8875-8a01b9672adb\") " pod="openstack/nova-cell1-db-create-94bx7" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.104311 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43b09009-a1b8-4f5f-8875-8a01b9672adb-operator-scripts\") pod \"nova-cell1-db-create-94bx7\" (UID: \"43b09009-a1b8-4f5f-8875-8a01b9672adb\") " pod="openstack/nova-cell1-db-create-94bx7" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.104348 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8121d8f-c473-4818-80f2-1335e44a7955-operator-scripts\") pod \"nova-cell0-3e19-account-create-update-bxrv4\" (UID: \"e8121d8f-c473-4818-80f2-1335e44a7955\") " pod="openstack/nova-cell0-3e19-account-create-update-bxrv4" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.105611 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43b09009-a1b8-4f5f-8875-8a01b9672adb-operator-scripts\") pod \"nova-cell1-db-create-94bx7\" (UID: \"43b09009-a1b8-4f5f-8875-8a01b9672adb\") " pod="openstack/nova-cell1-db-create-94bx7" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.121902 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8121d8f-c473-4818-80f2-1335e44a7955-operator-scripts\") pod \"nova-cell0-3e19-account-create-update-bxrv4\" (UID: \"e8121d8f-c473-4818-80f2-1335e44a7955\") " pod="openstack/nova-cell0-3e19-account-create-update-bxrv4" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.130904 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnvqb\" (UniqueName: \"kubernetes.io/projected/e8121d8f-c473-4818-80f2-1335e44a7955-kube-api-access-vnvqb\") pod \"nova-cell0-3e19-account-create-update-bxrv4\" (UID: \"e8121d8f-c473-4818-80f2-1335e44a7955\") " pod="openstack/nova-cell0-3e19-account-create-update-bxrv4" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.134469 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6ggj\" (UniqueName: \"kubernetes.io/projected/43b09009-a1b8-4f5f-8875-8a01b9672adb-kube-api-access-s6ggj\") pod \"nova-cell1-db-create-94bx7\" (UID: \"43b09009-a1b8-4f5f-8875-8a01b9672adb\") " pod="openstack/nova-cell1-db-create-94bx7" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.164087 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-9989-account-create-update-f5ft6"] Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.168129 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9989-account-create-update-f5ft6" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.172143 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.189674 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9989-account-create-update-f5ft6"] Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.223927 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/007dfd5a-635a-42ff-b733-1022fe21b6e4-operator-scripts\") pod \"nova-cell1-9989-account-create-update-f5ft6\" (UID: \"007dfd5a-635a-42ff-b733-1022fe21b6e4\") " pod="openstack/nova-cell1-9989-account-create-update-f5ft6" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.224043 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ckx6\" (UniqueName: \"kubernetes.io/projected/007dfd5a-635a-42ff-b733-1022fe21b6e4-kube-api-access-2ckx6\") pod \"nova-cell1-9989-account-create-update-f5ft6\" (UID: \"007dfd5a-635a-42ff-b733-1022fe21b6e4\") " pod="openstack/nova-cell1-9989-account-create-update-f5ft6" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.239601 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-94bx7" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.325424 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/007dfd5a-635a-42ff-b733-1022fe21b6e4-operator-scripts\") pod \"nova-cell1-9989-account-create-update-f5ft6\" (UID: \"007dfd5a-635a-42ff-b733-1022fe21b6e4\") " pod="openstack/nova-cell1-9989-account-create-update-f5ft6" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.325732 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ckx6\" (UniqueName: \"kubernetes.io/projected/007dfd5a-635a-42ff-b733-1022fe21b6e4-kube-api-access-2ckx6\") pod \"nova-cell1-9989-account-create-update-f5ft6\" (UID: \"007dfd5a-635a-42ff-b733-1022fe21b6e4\") " pod="openstack/nova-cell1-9989-account-create-update-f5ft6" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.326648 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/007dfd5a-635a-42ff-b733-1022fe21b6e4-operator-scripts\") pod \"nova-cell1-9989-account-create-update-f5ft6\" (UID: \"007dfd5a-635a-42ff-b733-1022fe21b6e4\") " pod="openstack/nova-cell1-9989-account-create-update-f5ft6" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.327714 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3e19-account-create-update-bxrv4" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.342885 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.354894 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5c6f7968ff-b5wb6" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.359361 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ckx6\" (UniqueName: \"kubernetes.io/projected/007dfd5a-635a-42ff-b733-1022fe21b6e4-kube-api-access-2ckx6\") pod \"nova-cell1-9989-account-create-update-f5ft6\" (UID: \"007dfd5a-635a-42ff-b733-1022fe21b6e4\") " pod="openstack/nova-cell1-9989-account-create-update-f5ft6" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.607111 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9989-account-create-update-f5ft6" Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.757018 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-msvcb"] Feb 14 11:28:19 crc kubenswrapper[4904]: I0214 11:28:19.796503 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-sjsmr"] Feb 14 11:28:20 crc kubenswrapper[4904]: I0214 11:28:20.021036 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-msvcb" event={"ID":"1d63d7a3-5e12-428e-a9dc-705b28836ca0","Type":"ContainerStarted","Data":"e68745e52a8b0544f082f8a55e68024779d5fff3e08bc482ba95571a9ab6b6f5"} Feb 14 11:28:20 crc kubenswrapper[4904]: I0214 11:28:20.023235 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-sjsmr" event={"ID":"60971f9e-6866-4f58-9686-0eb77268650a","Type":"ContainerStarted","Data":"aa8b74777f761bba1e2040414cb14cf00abb3bf3d9f1bc11e9188444ab3134e6"} Feb 14 11:28:20 crc kubenswrapper[4904]: I0214 11:28:20.026020 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"562e6e60-d241-4e05-98da-80b3c449919e","Type":"ContainerStarted","Data":"7ca9194e4d58b9e1a58c9b47437acc6619b0eec3873dabc06913d8f1b299657f"} Feb 14 11:28:20 crc kubenswrapper[4904]: I0214 11:28:20.168056 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-94bx7"] Feb 14 11:28:20 crc kubenswrapper[4904]: I0214 11:28:20.184300 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-65a0-account-create-update-lz527"] Feb 14 11:28:20 crc kubenswrapper[4904]: I0214 11:28:20.203106 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-3e19-account-create-update-bxrv4"] Feb 14 11:28:20 crc kubenswrapper[4904]: I0214 11:28:20.485661 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9989-account-create-update-f5ft6"] Feb 14 11:28:20 crc kubenswrapper[4904]: W0214 11:28:20.503421 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod007dfd5a_635a_42ff_b733_1022fe21b6e4.slice/crio-a7a090c8f5394460852951c6848f0dbc44924f2b3275a2788b0a6bf0615fa40a WatchSource:0}: Error finding container a7a090c8f5394460852951c6848f0dbc44924f2b3275a2788b0a6bf0615fa40a: Status 404 returned error can't find the container with id a7a090c8f5394460852951c6848f0dbc44924f2b3275a2788b0a6bf0615fa40a Feb 14 11:28:20 crc kubenswrapper[4904]: I0214 11:28:20.978449 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:21 crc kubenswrapper[4904]: I0214 11:28:21.060322 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-msvcb" event={"ID":"1d63d7a3-5e12-428e-a9dc-705b28836ca0","Type":"ContainerStarted","Data":"7bbe956539df1f0dd3029213f6239b45e4c549f580c63afb238146cbf2714a48"} Feb 14 11:28:21 crc kubenswrapper[4904]: I0214 11:28:21.073668 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9989-account-create-update-f5ft6" event={"ID":"007dfd5a-635a-42ff-b733-1022fe21b6e4","Type":"ContainerStarted","Data":"5b00d53d5a2e3c969cf00ca830a19f3265c27bd06c9f6bd11562172db8231133"} Feb 14 11:28:21 crc kubenswrapper[4904]: I0214 11:28:21.073710 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9989-account-create-update-f5ft6" event={"ID":"007dfd5a-635a-42ff-b733-1022fe21b6e4","Type":"ContainerStarted","Data":"a7a090c8f5394460852951c6848f0dbc44924f2b3275a2788b0a6bf0615fa40a"} Feb 14 11:28:21 crc kubenswrapper[4904]: I0214 11:28:21.077098 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-msvcb" podStartSLOduration=3.07708225 podStartE2EDuration="3.07708225s" podCreationTimestamp="2026-02-14 11:28:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:28:21.076222456 +0000 UTC m=+1091.888987117" watchObservedRunningTime="2026-02-14 11:28:21.07708225 +0000 UTC m=+1091.889846911" Feb 14 11:28:21 crc kubenswrapper[4904]: I0214 11:28:21.086649 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-94bx7" event={"ID":"43b09009-a1b8-4f5f-8875-8a01b9672adb","Type":"ContainerStarted","Data":"38d86ea826aaa0b0c338baa908f8e2eec8ab8645a24eb5da05a73e1148880428"} Feb 14 11:28:21 crc kubenswrapper[4904]: I0214 11:28:21.086686 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-94bx7" event={"ID":"43b09009-a1b8-4f5f-8875-8a01b9672adb","Type":"ContainerStarted","Data":"9404c83e11164128c7cf94278e9f4ce655a720c063db521f9fc8c72d3363cb6b"} Feb 14 11:28:21 crc kubenswrapper[4904]: I0214 11:28:21.101556 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-9989-account-create-update-f5ft6" podStartSLOduration=2.101537532 podStartE2EDuration="2.101537532s" podCreationTimestamp="2026-02-14 11:28:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:28:21.094104678 +0000 UTC m=+1091.906869339" watchObservedRunningTime="2026-02-14 11:28:21.101537532 +0000 UTC m=+1091.914302193" Feb 14 11:28:21 crc kubenswrapper[4904]: I0214 11:28:21.108124 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-sjsmr" event={"ID":"60971f9e-6866-4f58-9686-0eb77268650a","Type":"ContainerStarted","Data":"a7c9eb679c798b42518db4d0d8280720eaad17b33ac18c43dd6196669663e7a4"} Feb 14 11:28:21 crc kubenswrapper[4904]: I0214 11:28:21.117027 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-65a0-account-create-update-lz527" event={"ID":"193a644c-698e-4ad6-a36d-8c19c8ef0d0a","Type":"ContainerStarted","Data":"5617a814dc9d8473995bd32f3185aa1c209505034d59ccfbb4cf88e183c7b9bb"} Feb 14 11:28:21 crc kubenswrapper[4904]: I0214 11:28:21.117069 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-65a0-account-create-update-lz527" event={"ID":"193a644c-698e-4ad6-a36d-8c19c8ef0d0a","Type":"ContainerStarted","Data":"aa2fcc568132645f88bab549381ec10bdf4653b396d16a3418bfe14eabf24ec2"} Feb 14 11:28:21 crc kubenswrapper[4904]: I0214 11:28:21.122471 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-94bx7" podStartSLOduration=3.122453147 podStartE2EDuration="3.122453147s" podCreationTimestamp="2026-02-14 11:28:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:28:21.122294333 +0000 UTC m=+1091.935058994" watchObservedRunningTime="2026-02-14 11:28:21.122453147 +0000 UTC m=+1091.935217808" Feb 14 11:28:21 crc kubenswrapper[4904]: I0214 11:28:21.131987 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3e19-account-create-update-bxrv4" event={"ID":"e8121d8f-c473-4818-80f2-1335e44a7955","Type":"ContainerStarted","Data":"f5d2afde37cb14f49005661ff25620cc8f7cb4912b6cff6c69c1cd7bad02af86"} Feb 14 11:28:21 crc kubenswrapper[4904]: I0214 11:28:21.132031 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3e19-account-create-update-bxrv4" event={"ID":"e8121d8f-c473-4818-80f2-1335e44a7955","Type":"ContainerStarted","Data":"587acce260114028ebb36feb870b1881bf8ad701653cfe5d5e0a410f5b89a0a6"} Feb 14 11:28:21 crc kubenswrapper[4904]: I0214 11:28:21.145982 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-65a0-account-create-update-lz527" podStartSLOduration=3.145861601 podStartE2EDuration="3.145861601s" podCreationTimestamp="2026-02-14 11:28:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:28:21.140221646 +0000 UTC m=+1091.952986307" watchObservedRunningTime="2026-02-14 11:28:21.145861601 +0000 UTC m=+1091.958626262" Feb 14 11:28:21 crc kubenswrapper[4904]: I0214 11:28:21.174280 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-sjsmr" podStartSLOduration=3.174263682 podStartE2EDuration="3.174263682s" podCreationTimestamp="2026-02-14 11:28:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:28:21.173414939 +0000 UTC m=+1091.986179600" watchObservedRunningTime="2026-02-14 11:28:21.174263682 +0000 UTC m=+1091.987028333" Feb 14 11:28:21 crc kubenswrapper[4904]: I0214 11:28:21.200403 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-3e19-account-create-update-bxrv4" podStartSLOduration=3.200386231 podStartE2EDuration="3.200386231s" podCreationTimestamp="2026-02-14 11:28:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:28:21.193797619 +0000 UTC m=+1092.006562270" watchObservedRunningTime="2026-02-14 11:28:21.200386231 +0000 UTC m=+1092.013150892" Feb 14 11:28:22 crc kubenswrapper[4904]: I0214 11:28:22.141375 4904 generic.go:334] "Generic (PLEG): container finished" podID="43b09009-a1b8-4f5f-8875-8a01b9672adb" containerID="38d86ea826aaa0b0c338baa908f8e2eec8ab8645a24eb5da05a73e1148880428" exitCode=0 Feb 14 11:28:22 crc kubenswrapper[4904]: I0214 11:28:22.141483 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-94bx7" event={"ID":"43b09009-a1b8-4f5f-8875-8a01b9672adb","Type":"ContainerDied","Data":"38d86ea826aaa0b0c338baa908f8e2eec8ab8645a24eb5da05a73e1148880428"} Feb 14 11:28:22 crc kubenswrapper[4904]: I0214 11:28:22.143629 4904 generic.go:334] "Generic (PLEG): container finished" podID="60971f9e-6866-4f58-9686-0eb77268650a" containerID="a7c9eb679c798b42518db4d0d8280720eaad17b33ac18c43dd6196669663e7a4" exitCode=0 Feb 14 11:28:22 crc kubenswrapper[4904]: I0214 11:28:22.143677 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-sjsmr" event={"ID":"60971f9e-6866-4f58-9686-0eb77268650a","Type":"ContainerDied","Data":"a7c9eb679c798b42518db4d0d8280720eaad17b33ac18c43dd6196669663e7a4"} Feb 14 11:28:22 crc kubenswrapper[4904]: I0214 11:28:22.146087 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"562e6e60-d241-4e05-98da-80b3c449919e","Type":"ContainerStarted","Data":"e396c0c884d14cb75048ea12fde21e216b5370eb8a474ce53bc125394956f921"} Feb 14 11:28:22 crc kubenswrapper[4904]: I0214 11:28:22.147925 4904 generic.go:334] "Generic (PLEG): container finished" podID="1d63d7a3-5e12-428e-a9dc-705b28836ca0" containerID="7bbe956539df1f0dd3029213f6239b45e4c549f580c63afb238146cbf2714a48" exitCode=0 Feb 14 11:28:22 crc kubenswrapper[4904]: I0214 11:28:22.148070 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-msvcb" event={"ID":"1d63d7a3-5e12-428e-a9dc-705b28836ca0","Type":"ContainerDied","Data":"7bbe956539df1f0dd3029213f6239b45e4c549f580c63afb238146cbf2714a48"} Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.131195 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.161940 4904 generic.go:334] "Generic (PLEG): container finished" podID="007dfd5a-635a-42ff-b733-1022fe21b6e4" containerID="5b00d53d5a2e3c969cf00ca830a19f3265c27bd06c9f6bd11562172db8231133" exitCode=0 Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.162012 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9989-account-create-update-f5ft6" event={"ID":"007dfd5a-635a-42ff-b733-1022fe21b6e4","Type":"ContainerDied","Data":"5b00d53d5a2e3c969cf00ca830a19f3265c27bd06c9f6bd11562172db8231133"} Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.164818 4904 generic.go:334] "Generic (PLEG): container finished" podID="193a644c-698e-4ad6-a36d-8c19c8ef0d0a" containerID="5617a814dc9d8473995bd32f3185aa1c209505034d59ccfbb4cf88e183c7b9bb" exitCode=0 Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.164894 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-65a0-account-create-update-lz527" event={"ID":"193a644c-698e-4ad6-a36d-8c19c8ef0d0a","Type":"ContainerDied","Data":"5617a814dc9d8473995bd32f3185aa1c209505034d59ccfbb4cf88e183c7b9bb"} Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.168097 4904 generic.go:334] "Generic (PLEG): container finished" podID="e8121d8f-c473-4818-80f2-1335e44a7955" containerID="f5d2afde37cb14f49005661ff25620cc8f7cb4912b6cff6c69c1cd7bad02af86" exitCode=0 Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.168155 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3e19-account-create-update-bxrv4" event={"ID":"e8121d8f-c473-4818-80f2-1335e44a7955","Type":"ContainerDied","Data":"f5d2afde37cb14f49005661ff25620cc8f7cb4912b6cff6c69c1cd7bad02af86"} Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.633750 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-sjsmr" Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.753817 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rngr5\" (UniqueName: \"kubernetes.io/projected/60971f9e-6866-4f58-9686-0eb77268650a-kube-api-access-rngr5\") pod \"60971f9e-6866-4f58-9686-0eb77268650a\" (UID: \"60971f9e-6866-4f58-9686-0eb77268650a\") " Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.753944 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60971f9e-6866-4f58-9686-0eb77268650a-operator-scripts\") pod \"60971f9e-6866-4f58-9686-0eb77268650a\" (UID: \"60971f9e-6866-4f58-9686-0eb77268650a\") " Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.754647 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60971f9e-6866-4f58-9686-0eb77268650a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "60971f9e-6866-4f58-9686-0eb77268650a" (UID: "60971f9e-6866-4f58-9686-0eb77268650a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.761648 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60971f9e-6866-4f58-9686-0eb77268650a-kube-api-access-rngr5" (OuterVolumeSpecName: "kube-api-access-rngr5") pod "60971f9e-6866-4f58-9686-0eb77268650a" (UID: "60971f9e-6866-4f58-9686-0eb77268650a"). InnerVolumeSpecName "kube-api-access-rngr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.776308 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-msvcb" Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.858218 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rngr5\" (UniqueName: \"kubernetes.io/projected/60971f9e-6866-4f58-9686-0eb77268650a-kube-api-access-rngr5\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.860923 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60971f9e-6866-4f58-9686-0eb77268650a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.932883 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-94bx7" Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.962527 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d63d7a3-5e12-428e-a9dc-705b28836ca0-operator-scripts\") pod \"1d63d7a3-5e12-428e-a9dc-705b28836ca0\" (UID: \"1d63d7a3-5e12-428e-a9dc-705b28836ca0\") " Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.962722 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwppd\" (UniqueName: \"kubernetes.io/projected/1d63d7a3-5e12-428e-a9dc-705b28836ca0-kube-api-access-mwppd\") pod \"1d63d7a3-5e12-428e-a9dc-705b28836ca0\" (UID: \"1d63d7a3-5e12-428e-a9dc-705b28836ca0\") " Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.964514 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d63d7a3-5e12-428e-a9dc-705b28836ca0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1d63d7a3-5e12-428e-a9dc-705b28836ca0" (UID: "1d63d7a3-5e12-428e-a9dc-705b28836ca0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:28:23 crc kubenswrapper[4904]: I0214 11:28:23.970065 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d63d7a3-5e12-428e-a9dc-705b28836ca0-kube-api-access-mwppd" (OuterVolumeSpecName: "kube-api-access-mwppd") pod "1d63d7a3-5e12-428e-a9dc-705b28836ca0" (UID: "1d63d7a3-5e12-428e-a9dc-705b28836ca0"). InnerVolumeSpecName "kube-api-access-mwppd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.032801 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.032870 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.066347 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6ggj\" (UniqueName: \"kubernetes.io/projected/43b09009-a1b8-4f5f-8875-8a01b9672adb-kube-api-access-s6ggj\") pod \"43b09009-a1b8-4f5f-8875-8a01b9672adb\" (UID: \"43b09009-a1b8-4f5f-8875-8a01b9672adb\") " Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.066502 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43b09009-a1b8-4f5f-8875-8a01b9672adb-operator-scripts\") pod \"43b09009-a1b8-4f5f-8875-8a01b9672adb\" (UID: \"43b09009-a1b8-4f5f-8875-8a01b9672adb\") " Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.067227 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwppd\" (UniqueName: \"kubernetes.io/projected/1d63d7a3-5e12-428e-a9dc-705b28836ca0-kube-api-access-mwppd\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.067250 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d63d7a3-5e12-428e-a9dc-705b28836ca0-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.068026 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43b09009-a1b8-4f5f-8875-8a01b9672adb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "43b09009-a1b8-4f5f-8875-8a01b9672adb" (UID: "43b09009-a1b8-4f5f-8875-8a01b9672adb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.076703 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43b09009-a1b8-4f5f-8875-8a01b9672adb-kube-api-access-s6ggj" (OuterVolumeSpecName: "kube-api-access-s6ggj") pod "43b09009-a1b8-4f5f-8875-8a01b9672adb" (UID: "43b09009-a1b8-4f5f-8875-8a01b9672adb"). InnerVolumeSpecName "kube-api-access-s6ggj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.095093 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.095169 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.118962 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.119035 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.170407 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6ggj\" (UniqueName: \"kubernetes.io/projected/43b09009-a1b8-4f5f-8875-8a01b9672adb-kube-api-access-s6ggj\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.170431 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43b09009-a1b8-4f5f-8875-8a01b9672adb-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.172699 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.175557 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.187987 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"562e6e60-d241-4e05-98da-80b3c449919e","Type":"ContainerStarted","Data":"d291064ffeecd50e3368f92b089d799eadc12d6b56696353ffc3968fd10d1c01"} Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.188315 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.188303 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="562e6e60-d241-4e05-98da-80b3c449919e" containerName="ceilometer-central-agent" containerID="cri-o://65653abe18fc0a6a7f40e2f85fafd80b5c9ac25b277b21ddc74f8fab2e346a63" gracePeriod=30 Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.188325 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="562e6e60-d241-4e05-98da-80b3c449919e" containerName="sg-core" containerID="cri-o://e396c0c884d14cb75048ea12fde21e216b5370eb8a474ce53bc125394956f921" gracePeriod=30 Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.188367 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="562e6e60-d241-4e05-98da-80b3c449919e" containerName="proxy-httpd" containerID="cri-o://d291064ffeecd50e3368f92b089d799eadc12d6b56696353ffc3968fd10d1c01" gracePeriod=30 Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.188441 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="562e6e60-d241-4e05-98da-80b3c449919e" containerName="ceilometer-notification-agent" containerID="cri-o://7ca9194e4d58b9e1a58c9b47437acc6619b0eec3873dabc06913d8f1b299657f" gracePeriod=30 Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.197974 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-msvcb" event={"ID":"1d63d7a3-5e12-428e-a9dc-705b28836ca0","Type":"ContainerDied","Data":"e68745e52a8b0544f082f8a55e68024779d5fff3e08bc482ba95571a9ab6b6f5"} Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.198020 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e68745e52a8b0544f082f8a55e68024779d5fff3e08bc482ba95571a9ab6b6f5" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.198091 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-msvcb" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.201306 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-94bx7" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.201308 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-94bx7" event={"ID":"43b09009-a1b8-4f5f-8875-8a01b9672adb","Type":"ContainerDied","Data":"9404c83e11164128c7cf94278e9f4ce655a720c063db521f9fc8c72d3363cb6b"} Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.201427 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9404c83e11164128c7cf94278e9f4ce655a720c063db521f9fc8c72d3363cb6b" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.204915 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-sjsmr" event={"ID":"60971f9e-6866-4f58-9686-0eb77268650a","Type":"ContainerDied","Data":"aa8b74777f761bba1e2040414cb14cf00abb3bf3d9f1bc11e9188444ab3134e6"} Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.205110 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa8b74777f761bba1e2040414cb14cf00abb3bf3d9f1bc11e9188444ab3134e6" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.205000 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-sjsmr" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.205358 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.205467 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.205513 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.205525 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.260230 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.694181204 podStartE2EDuration="8.26021079s" podCreationTimestamp="2026-02-14 11:28:16 +0000 UTC" firstStartedPulling="2026-02-14 11:28:17.471993426 +0000 UTC m=+1088.284758087" lastFinishedPulling="2026-02-14 11:28:23.038023022 +0000 UTC m=+1093.850787673" observedRunningTime="2026-02-14 11:28:24.231859651 +0000 UTC m=+1095.044624312" watchObservedRunningTime="2026-02-14 11:28:24.26021079 +0000 UTC m=+1095.072975451" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.756799 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-65a0-account-create-update-lz527" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.774075 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3e19-account-create-update-bxrv4" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.806294 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9989-account-create-update-f5ft6" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.886768 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/193a644c-698e-4ad6-a36d-8c19c8ef0d0a-operator-scripts\") pod \"193a644c-698e-4ad6-a36d-8c19c8ef0d0a\" (UID: \"193a644c-698e-4ad6-a36d-8c19c8ef0d0a\") " Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.887124 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4cmz\" (UniqueName: \"kubernetes.io/projected/193a644c-698e-4ad6-a36d-8c19c8ef0d0a-kube-api-access-m4cmz\") pod \"193a644c-698e-4ad6-a36d-8c19c8ef0d0a\" (UID: \"193a644c-698e-4ad6-a36d-8c19c8ef0d0a\") " Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.887176 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnvqb\" (UniqueName: \"kubernetes.io/projected/e8121d8f-c473-4818-80f2-1335e44a7955-kube-api-access-vnvqb\") pod \"e8121d8f-c473-4818-80f2-1335e44a7955\" (UID: \"e8121d8f-c473-4818-80f2-1335e44a7955\") " Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.887376 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ckx6\" (UniqueName: \"kubernetes.io/projected/007dfd5a-635a-42ff-b733-1022fe21b6e4-kube-api-access-2ckx6\") pod \"007dfd5a-635a-42ff-b733-1022fe21b6e4\" (UID: \"007dfd5a-635a-42ff-b733-1022fe21b6e4\") " Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.887487 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/007dfd5a-635a-42ff-b733-1022fe21b6e4-operator-scripts\") pod \"007dfd5a-635a-42ff-b733-1022fe21b6e4\" (UID: \"007dfd5a-635a-42ff-b733-1022fe21b6e4\") " Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.887507 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8121d8f-c473-4818-80f2-1335e44a7955-operator-scripts\") pod \"e8121d8f-c473-4818-80f2-1335e44a7955\" (UID: \"e8121d8f-c473-4818-80f2-1335e44a7955\") " Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.889435 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/007dfd5a-635a-42ff-b733-1022fe21b6e4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "007dfd5a-635a-42ff-b733-1022fe21b6e4" (UID: "007dfd5a-635a-42ff-b733-1022fe21b6e4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.889956 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/193a644c-698e-4ad6-a36d-8c19c8ef0d0a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "193a644c-698e-4ad6-a36d-8c19c8ef0d0a" (UID: "193a644c-698e-4ad6-a36d-8c19c8ef0d0a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.890275 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8121d8f-c473-4818-80f2-1335e44a7955-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e8121d8f-c473-4818-80f2-1335e44a7955" (UID: "e8121d8f-c473-4818-80f2-1335e44a7955"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.896614 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/007dfd5a-635a-42ff-b733-1022fe21b6e4-kube-api-access-2ckx6" (OuterVolumeSpecName: "kube-api-access-2ckx6") pod "007dfd5a-635a-42ff-b733-1022fe21b6e4" (UID: "007dfd5a-635a-42ff-b733-1022fe21b6e4"). InnerVolumeSpecName "kube-api-access-2ckx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.902979 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/193a644c-698e-4ad6-a36d-8c19c8ef0d0a-kube-api-access-m4cmz" (OuterVolumeSpecName: "kube-api-access-m4cmz") pod "193a644c-698e-4ad6-a36d-8c19c8ef0d0a" (UID: "193a644c-698e-4ad6-a36d-8c19c8ef0d0a"). InnerVolumeSpecName "kube-api-access-m4cmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.925861 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8121d8f-c473-4818-80f2-1335e44a7955-kube-api-access-vnvqb" (OuterVolumeSpecName: "kube-api-access-vnvqb") pod "e8121d8f-c473-4818-80f2-1335e44a7955" (UID: "e8121d8f-c473-4818-80f2-1335e44a7955"). InnerVolumeSpecName "kube-api-access-vnvqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.991398 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ckx6\" (UniqueName: \"kubernetes.io/projected/007dfd5a-635a-42ff-b733-1022fe21b6e4-kube-api-access-2ckx6\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.991434 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/007dfd5a-635a-42ff-b733-1022fe21b6e4-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.991443 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8121d8f-c473-4818-80f2-1335e44a7955-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.991452 4904 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/193a644c-698e-4ad6-a36d-8c19c8ef0d0a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.991461 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4cmz\" (UniqueName: \"kubernetes.io/projected/193a644c-698e-4ad6-a36d-8c19c8ef0d0a-kube-api-access-m4cmz\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:24 crc kubenswrapper[4904]: I0214 11:28:24.991470 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnvqb\" (UniqueName: \"kubernetes.io/projected/e8121d8f-c473-4818-80f2-1335e44a7955-kube-api-access-vnvqb\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:25 crc kubenswrapper[4904]: I0214 11:28:25.217353 4904 generic.go:334] "Generic (PLEG): container finished" podID="562e6e60-d241-4e05-98da-80b3c449919e" containerID="d291064ffeecd50e3368f92b089d799eadc12d6b56696353ffc3968fd10d1c01" exitCode=0 Feb 14 11:28:25 crc kubenswrapper[4904]: I0214 11:28:25.217391 4904 generic.go:334] "Generic (PLEG): container finished" podID="562e6e60-d241-4e05-98da-80b3c449919e" containerID="e396c0c884d14cb75048ea12fde21e216b5370eb8a474ce53bc125394956f921" exitCode=2 Feb 14 11:28:25 crc kubenswrapper[4904]: I0214 11:28:25.217403 4904 generic.go:334] "Generic (PLEG): container finished" podID="562e6e60-d241-4e05-98da-80b3c449919e" containerID="7ca9194e4d58b9e1a58c9b47437acc6619b0eec3873dabc06913d8f1b299657f" exitCode=0 Feb 14 11:28:25 crc kubenswrapper[4904]: I0214 11:28:25.217509 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"562e6e60-d241-4e05-98da-80b3c449919e","Type":"ContainerDied","Data":"d291064ffeecd50e3368f92b089d799eadc12d6b56696353ffc3968fd10d1c01"} Feb 14 11:28:25 crc kubenswrapper[4904]: I0214 11:28:25.217543 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"562e6e60-d241-4e05-98da-80b3c449919e","Type":"ContainerDied","Data":"e396c0c884d14cb75048ea12fde21e216b5370eb8a474ce53bc125394956f921"} Feb 14 11:28:25 crc kubenswrapper[4904]: I0214 11:28:25.217556 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"562e6e60-d241-4e05-98da-80b3c449919e","Type":"ContainerDied","Data":"7ca9194e4d58b9e1a58c9b47437acc6619b0eec3873dabc06913d8f1b299657f"} Feb 14 11:28:25 crc kubenswrapper[4904]: I0214 11:28:25.219927 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9989-account-create-update-f5ft6" event={"ID":"007dfd5a-635a-42ff-b733-1022fe21b6e4","Type":"ContainerDied","Data":"a7a090c8f5394460852951c6848f0dbc44924f2b3275a2788b0a6bf0615fa40a"} Feb 14 11:28:25 crc kubenswrapper[4904]: I0214 11:28:25.219974 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9989-account-create-update-f5ft6" Feb 14 11:28:25 crc kubenswrapper[4904]: I0214 11:28:25.219991 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7a090c8f5394460852951c6848f0dbc44924f2b3275a2788b0a6bf0615fa40a" Feb 14 11:28:25 crc kubenswrapper[4904]: I0214 11:28:25.222284 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-65a0-account-create-update-lz527" event={"ID":"193a644c-698e-4ad6-a36d-8c19c8ef0d0a","Type":"ContainerDied","Data":"aa2fcc568132645f88bab549381ec10bdf4653b396d16a3418bfe14eabf24ec2"} Feb 14 11:28:25 crc kubenswrapper[4904]: I0214 11:28:25.222300 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-65a0-account-create-update-lz527" Feb 14 11:28:25 crc kubenswrapper[4904]: I0214 11:28:25.222312 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa2fcc568132645f88bab549381ec10bdf4653b396d16a3418bfe14eabf24ec2" Feb 14 11:28:25 crc kubenswrapper[4904]: I0214 11:28:25.229061 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3e19-account-create-update-bxrv4" Feb 14 11:28:25 crc kubenswrapper[4904]: I0214 11:28:25.229502 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3e19-account-create-update-bxrv4" event={"ID":"e8121d8f-c473-4818-80f2-1335e44a7955","Type":"ContainerDied","Data":"587acce260114028ebb36feb870b1881bf8ad701653cfe5d5e0a410f5b89a0a6"} Feb 14 11:28:25 crc kubenswrapper[4904]: I0214 11:28:25.229542 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="587acce260114028ebb36feb870b1881bf8ad701653cfe5d5e0a410f5b89a0a6" Feb 14 11:28:27 crc kubenswrapper[4904]: I0214 11:28:27.495586 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 14 11:28:27 crc kubenswrapper[4904]: I0214 11:28:27.496175 4904 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 14 11:28:27 crc kubenswrapper[4904]: I0214 11:28:27.505199 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 14 11:28:27 crc kubenswrapper[4904]: I0214 11:28:27.525898 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 14 11:28:27 crc kubenswrapper[4904]: I0214 11:28:27.526085 4904 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 14 11:28:27 crc kubenswrapper[4904]: I0214 11:28:27.830530 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 14 11:28:28 crc kubenswrapper[4904]: I0214 11:28:28.783053 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6bc9d549b8-hdn6n" podUID="cfb252c0-ffb2-4d87-9c4e-b56cc2070d71" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 14 11:28:28 crc kubenswrapper[4904]: I0214 11:28:28.950937 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-687657b496-tz596" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.526665 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xdnmn"] Feb 14 11:28:29 crc kubenswrapper[4904]: E0214 11:28:29.528556 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8121d8f-c473-4818-80f2-1335e44a7955" containerName="mariadb-account-create-update" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.528694 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8121d8f-c473-4818-80f2-1335e44a7955" containerName="mariadb-account-create-update" Feb 14 11:28:29 crc kubenswrapper[4904]: E0214 11:28:29.528791 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="193a644c-698e-4ad6-a36d-8c19c8ef0d0a" containerName="mariadb-account-create-update" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.530935 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="193a644c-698e-4ad6-a36d-8c19c8ef0d0a" containerName="mariadb-account-create-update" Feb 14 11:28:29 crc kubenswrapper[4904]: E0214 11:28:29.531045 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="007dfd5a-635a-42ff-b733-1022fe21b6e4" containerName="mariadb-account-create-update" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.531130 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="007dfd5a-635a-42ff-b733-1022fe21b6e4" containerName="mariadb-account-create-update" Feb 14 11:28:29 crc kubenswrapper[4904]: E0214 11:28:29.531219 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d63d7a3-5e12-428e-a9dc-705b28836ca0" containerName="mariadb-database-create" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.531299 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d63d7a3-5e12-428e-a9dc-705b28836ca0" containerName="mariadb-database-create" Feb 14 11:28:29 crc kubenswrapper[4904]: E0214 11:28:29.531446 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43b09009-a1b8-4f5f-8875-8a01b9672adb" containerName="mariadb-database-create" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.531577 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="43b09009-a1b8-4f5f-8875-8a01b9672adb" containerName="mariadb-database-create" Feb 14 11:28:29 crc kubenswrapper[4904]: E0214 11:28:29.531669 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60971f9e-6866-4f58-9686-0eb77268650a" containerName="mariadb-database-create" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.531757 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="60971f9e-6866-4f58-9686-0eb77268650a" containerName="mariadb-database-create" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.532180 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d63d7a3-5e12-428e-a9dc-705b28836ca0" containerName="mariadb-database-create" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.532291 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="60971f9e-6866-4f58-9686-0eb77268650a" containerName="mariadb-database-create" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.532448 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8121d8f-c473-4818-80f2-1335e44a7955" containerName="mariadb-account-create-update" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.532531 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="43b09009-a1b8-4f5f-8875-8a01b9672adb" containerName="mariadb-database-create" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.532614 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="193a644c-698e-4ad6-a36d-8c19c8ef0d0a" containerName="mariadb-account-create-update" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.532699 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="007dfd5a-635a-42ff-b733-1022fe21b6e4" containerName="mariadb-account-create-update" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.533518 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xdnmn" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.537204 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.538555 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-pl6lw" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.538591 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.540534 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xdnmn"] Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.684049 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6971dbc2-c11e-44d1-b63a-e0487f339183-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xdnmn\" (UID: \"6971dbc2-c11e-44d1-b63a-e0487f339183\") " pod="openstack/nova-cell0-conductor-db-sync-xdnmn" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.684156 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6971dbc2-c11e-44d1-b63a-e0487f339183-config-data\") pod \"nova-cell0-conductor-db-sync-xdnmn\" (UID: \"6971dbc2-c11e-44d1-b63a-e0487f339183\") " pod="openstack/nova-cell0-conductor-db-sync-xdnmn" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.684205 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6971dbc2-c11e-44d1-b63a-e0487f339183-scripts\") pod \"nova-cell0-conductor-db-sync-xdnmn\" (UID: \"6971dbc2-c11e-44d1-b63a-e0487f339183\") " pod="openstack/nova-cell0-conductor-db-sync-xdnmn" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.684245 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmjfp\" (UniqueName: \"kubernetes.io/projected/6971dbc2-c11e-44d1-b63a-e0487f339183-kube-api-access-nmjfp\") pod \"nova-cell0-conductor-db-sync-xdnmn\" (UID: \"6971dbc2-c11e-44d1-b63a-e0487f339183\") " pod="openstack/nova-cell0-conductor-db-sync-xdnmn" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.785268 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6971dbc2-c11e-44d1-b63a-e0487f339183-scripts\") pod \"nova-cell0-conductor-db-sync-xdnmn\" (UID: \"6971dbc2-c11e-44d1-b63a-e0487f339183\") " pod="openstack/nova-cell0-conductor-db-sync-xdnmn" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.785341 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmjfp\" (UniqueName: \"kubernetes.io/projected/6971dbc2-c11e-44d1-b63a-e0487f339183-kube-api-access-nmjfp\") pod \"nova-cell0-conductor-db-sync-xdnmn\" (UID: \"6971dbc2-c11e-44d1-b63a-e0487f339183\") " pod="openstack/nova-cell0-conductor-db-sync-xdnmn" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.785396 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6971dbc2-c11e-44d1-b63a-e0487f339183-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xdnmn\" (UID: \"6971dbc2-c11e-44d1-b63a-e0487f339183\") " pod="openstack/nova-cell0-conductor-db-sync-xdnmn" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.785456 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6971dbc2-c11e-44d1-b63a-e0487f339183-config-data\") pod \"nova-cell0-conductor-db-sync-xdnmn\" (UID: \"6971dbc2-c11e-44d1-b63a-e0487f339183\") " pod="openstack/nova-cell0-conductor-db-sync-xdnmn" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.791049 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6971dbc2-c11e-44d1-b63a-e0487f339183-scripts\") pod \"nova-cell0-conductor-db-sync-xdnmn\" (UID: \"6971dbc2-c11e-44d1-b63a-e0487f339183\") " pod="openstack/nova-cell0-conductor-db-sync-xdnmn" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.792528 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6971dbc2-c11e-44d1-b63a-e0487f339183-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xdnmn\" (UID: \"6971dbc2-c11e-44d1-b63a-e0487f339183\") " pod="openstack/nova-cell0-conductor-db-sync-xdnmn" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.794446 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6971dbc2-c11e-44d1-b63a-e0487f339183-config-data\") pod \"nova-cell0-conductor-db-sync-xdnmn\" (UID: \"6971dbc2-c11e-44d1-b63a-e0487f339183\") " pod="openstack/nova-cell0-conductor-db-sync-xdnmn" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.817975 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmjfp\" (UniqueName: \"kubernetes.io/projected/6971dbc2-c11e-44d1-b63a-e0487f339183-kube-api-access-nmjfp\") pod \"nova-cell0-conductor-db-sync-xdnmn\" (UID: \"6971dbc2-c11e-44d1-b63a-e0487f339183\") " pod="openstack/nova-cell0-conductor-db-sync-xdnmn" Feb 14 11:28:29 crc kubenswrapper[4904]: I0214 11:28:29.867722 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xdnmn" Feb 14 11:28:30 crc kubenswrapper[4904]: W0214 11:28:30.441279 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6971dbc2_c11e_44d1_b63a_e0487f339183.slice/crio-9bcb8f260ae0e4278935c16f871cb009260ad6ba0d37d36f7c0473ae5984437d WatchSource:0}: Error finding container 9bcb8f260ae0e4278935c16f871cb009260ad6ba0d37d36f7c0473ae5984437d: Status 404 returned error can't find the container with id 9bcb8f260ae0e4278935c16f871cb009260ad6ba0d37d36f7c0473ae5984437d Feb 14 11:28:30 crc kubenswrapper[4904]: I0214 11:28:30.441356 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xdnmn"] Feb 14 11:28:31 crc kubenswrapper[4904]: I0214 11:28:31.287649 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xdnmn" event={"ID":"6971dbc2-c11e-44d1-b63a-e0487f339183","Type":"ContainerStarted","Data":"9bcb8f260ae0e4278935c16f871cb009260ad6ba0d37d36f7c0473ae5984437d"} Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.706630 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.845927 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-config-data\") pod \"562e6e60-d241-4e05-98da-80b3c449919e\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.846188 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/562e6e60-d241-4e05-98da-80b3c449919e-log-httpd\") pod \"562e6e60-d241-4e05-98da-80b3c449919e\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.846233 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-sg-core-conf-yaml\") pod \"562e6e60-d241-4e05-98da-80b3c449919e\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.846267 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5k6xq\" (UniqueName: \"kubernetes.io/projected/562e6e60-d241-4e05-98da-80b3c449919e-kube-api-access-5k6xq\") pod \"562e6e60-d241-4e05-98da-80b3c449919e\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.846286 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-scripts\") pod \"562e6e60-d241-4e05-98da-80b3c449919e\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.846306 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-combined-ca-bundle\") pod \"562e6e60-d241-4e05-98da-80b3c449919e\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.846325 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/562e6e60-d241-4e05-98da-80b3c449919e-run-httpd\") pod \"562e6e60-d241-4e05-98da-80b3c449919e\" (UID: \"562e6e60-d241-4e05-98da-80b3c449919e\") " Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.846663 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/562e6e60-d241-4e05-98da-80b3c449919e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "562e6e60-d241-4e05-98da-80b3c449919e" (UID: "562e6e60-d241-4e05-98da-80b3c449919e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.846993 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/562e6e60-d241-4e05-98da-80b3c449919e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "562e6e60-d241-4e05-98da-80b3c449919e" (UID: "562e6e60-d241-4e05-98da-80b3c449919e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.878100 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-scripts" (OuterVolumeSpecName: "scripts") pod "562e6e60-d241-4e05-98da-80b3c449919e" (UID: "562e6e60-d241-4e05-98da-80b3c449919e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.881236 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/562e6e60-d241-4e05-98da-80b3c449919e-kube-api-access-5k6xq" (OuterVolumeSpecName: "kube-api-access-5k6xq") pod "562e6e60-d241-4e05-98da-80b3c449919e" (UID: "562e6e60-d241-4e05-98da-80b3c449919e"). InnerVolumeSpecName "kube-api-access-5k6xq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.943843 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "562e6e60-d241-4e05-98da-80b3c449919e" (UID: "562e6e60-d241-4e05-98da-80b3c449919e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.948584 4904 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/562e6e60-d241-4e05-98da-80b3c449919e-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.948614 4904 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.948625 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5k6xq\" (UniqueName: \"kubernetes.io/projected/562e6e60-d241-4e05-98da-80b3c449919e-kube-api-access-5k6xq\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.948633 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.948643 4904 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/562e6e60-d241-4e05-98da-80b3c449919e-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.975048 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "562e6e60-d241-4e05-98da-80b3c449919e" (UID: "562e6e60-d241-4e05-98da-80b3c449919e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:32 crc kubenswrapper[4904]: I0214 11:28:32.990543 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-config-data" (OuterVolumeSpecName: "config-data") pod "562e6e60-d241-4e05-98da-80b3c449919e" (UID: "562e6e60-d241-4e05-98da-80b3c449919e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.050059 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.050090 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/562e6e60-d241-4e05-98da-80b3c449919e-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.336360 4904 generic.go:334] "Generic (PLEG): container finished" podID="562e6e60-d241-4e05-98da-80b3c449919e" containerID="65653abe18fc0a6a7f40e2f85fafd80b5c9ac25b277b21ddc74f8fab2e346a63" exitCode=0 Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.336402 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"562e6e60-d241-4e05-98da-80b3c449919e","Type":"ContainerDied","Data":"65653abe18fc0a6a7f40e2f85fafd80b5c9ac25b277b21ddc74f8fab2e346a63"} Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.336417 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.336432 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"562e6e60-d241-4e05-98da-80b3c449919e","Type":"ContainerDied","Data":"9fe360eee0479af31170fbfb61585e3cdc19ab77ac99961d6e707232b0539953"} Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.336475 4904 scope.go:117] "RemoveContainer" containerID="d291064ffeecd50e3368f92b089d799eadc12d6b56696353ffc3968fd10d1c01" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.373650 4904 scope.go:117] "RemoveContainer" containerID="e396c0c884d14cb75048ea12fde21e216b5370eb8a474ce53bc125394956f921" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.402443 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.419229 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.436099 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:33 crc kubenswrapper[4904]: E0214 11:28:33.436516 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="562e6e60-d241-4e05-98da-80b3c449919e" containerName="sg-core" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.436533 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="562e6e60-d241-4e05-98da-80b3c449919e" containerName="sg-core" Feb 14 11:28:33 crc kubenswrapper[4904]: E0214 11:28:33.436544 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="562e6e60-d241-4e05-98da-80b3c449919e" containerName="ceilometer-central-agent" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.436551 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="562e6e60-d241-4e05-98da-80b3c449919e" containerName="ceilometer-central-agent" Feb 14 11:28:33 crc kubenswrapper[4904]: E0214 11:28:33.436565 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="562e6e60-d241-4e05-98da-80b3c449919e" containerName="ceilometer-notification-agent" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.436571 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="562e6e60-d241-4e05-98da-80b3c449919e" containerName="ceilometer-notification-agent" Feb 14 11:28:33 crc kubenswrapper[4904]: E0214 11:28:33.436580 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="562e6e60-d241-4e05-98da-80b3c449919e" containerName="proxy-httpd" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.436585 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="562e6e60-d241-4e05-98da-80b3c449919e" containerName="proxy-httpd" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.436786 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="562e6e60-d241-4e05-98da-80b3c449919e" containerName="proxy-httpd" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.436800 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="562e6e60-d241-4e05-98da-80b3c449919e" containerName="ceilometer-notification-agent" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.436812 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="562e6e60-d241-4e05-98da-80b3c449919e" containerName="ceilometer-central-agent" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.436822 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="562e6e60-d241-4e05-98da-80b3c449919e" containerName="sg-core" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.438367 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.439856 4904 scope.go:117] "RemoveContainer" containerID="7ca9194e4d58b9e1a58c9b47437acc6619b0eec3873dabc06913d8f1b299657f" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.442363 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.443020 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.443272 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.503526 4904 scope.go:117] "RemoveContainer" containerID="65653abe18fc0a6a7f40e2f85fafd80b5c9ac25b277b21ddc74f8fab2e346a63" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.522427 4904 scope.go:117] "RemoveContainer" containerID="d291064ffeecd50e3368f92b089d799eadc12d6b56696353ffc3968fd10d1c01" Feb 14 11:28:33 crc kubenswrapper[4904]: E0214 11:28:33.522906 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d291064ffeecd50e3368f92b089d799eadc12d6b56696353ffc3968fd10d1c01\": container with ID starting with d291064ffeecd50e3368f92b089d799eadc12d6b56696353ffc3968fd10d1c01 not found: ID does not exist" containerID="d291064ffeecd50e3368f92b089d799eadc12d6b56696353ffc3968fd10d1c01" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.522948 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d291064ffeecd50e3368f92b089d799eadc12d6b56696353ffc3968fd10d1c01"} err="failed to get container status \"d291064ffeecd50e3368f92b089d799eadc12d6b56696353ffc3968fd10d1c01\": rpc error: code = NotFound desc = could not find container \"d291064ffeecd50e3368f92b089d799eadc12d6b56696353ffc3968fd10d1c01\": container with ID starting with d291064ffeecd50e3368f92b089d799eadc12d6b56696353ffc3968fd10d1c01 not found: ID does not exist" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.522975 4904 scope.go:117] "RemoveContainer" containerID="e396c0c884d14cb75048ea12fde21e216b5370eb8a474ce53bc125394956f921" Feb 14 11:28:33 crc kubenswrapper[4904]: E0214 11:28:33.525073 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e396c0c884d14cb75048ea12fde21e216b5370eb8a474ce53bc125394956f921\": container with ID starting with e396c0c884d14cb75048ea12fde21e216b5370eb8a474ce53bc125394956f921 not found: ID does not exist" containerID="e396c0c884d14cb75048ea12fde21e216b5370eb8a474ce53bc125394956f921" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.525115 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e396c0c884d14cb75048ea12fde21e216b5370eb8a474ce53bc125394956f921"} err="failed to get container status \"e396c0c884d14cb75048ea12fde21e216b5370eb8a474ce53bc125394956f921\": rpc error: code = NotFound desc = could not find container \"e396c0c884d14cb75048ea12fde21e216b5370eb8a474ce53bc125394956f921\": container with ID starting with e396c0c884d14cb75048ea12fde21e216b5370eb8a474ce53bc125394956f921 not found: ID does not exist" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.525141 4904 scope.go:117] "RemoveContainer" containerID="7ca9194e4d58b9e1a58c9b47437acc6619b0eec3873dabc06913d8f1b299657f" Feb 14 11:28:33 crc kubenswrapper[4904]: E0214 11:28:33.525509 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ca9194e4d58b9e1a58c9b47437acc6619b0eec3873dabc06913d8f1b299657f\": container with ID starting with 7ca9194e4d58b9e1a58c9b47437acc6619b0eec3873dabc06913d8f1b299657f not found: ID does not exist" containerID="7ca9194e4d58b9e1a58c9b47437acc6619b0eec3873dabc06913d8f1b299657f" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.525572 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ca9194e4d58b9e1a58c9b47437acc6619b0eec3873dabc06913d8f1b299657f"} err="failed to get container status \"7ca9194e4d58b9e1a58c9b47437acc6619b0eec3873dabc06913d8f1b299657f\": rpc error: code = NotFound desc = could not find container \"7ca9194e4d58b9e1a58c9b47437acc6619b0eec3873dabc06913d8f1b299657f\": container with ID starting with 7ca9194e4d58b9e1a58c9b47437acc6619b0eec3873dabc06913d8f1b299657f not found: ID does not exist" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.525598 4904 scope.go:117] "RemoveContainer" containerID="65653abe18fc0a6a7f40e2f85fafd80b5c9ac25b277b21ddc74f8fab2e346a63" Feb 14 11:28:33 crc kubenswrapper[4904]: E0214 11:28:33.526060 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65653abe18fc0a6a7f40e2f85fafd80b5c9ac25b277b21ddc74f8fab2e346a63\": container with ID starting with 65653abe18fc0a6a7f40e2f85fafd80b5c9ac25b277b21ddc74f8fab2e346a63 not found: ID does not exist" containerID="65653abe18fc0a6a7f40e2f85fafd80b5c9ac25b277b21ddc74f8fab2e346a63" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.526084 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65653abe18fc0a6a7f40e2f85fafd80b5c9ac25b277b21ddc74f8fab2e346a63"} err="failed to get container status \"65653abe18fc0a6a7f40e2f85fafd80b5c9ac25b277b21ddc74f8fab2e346a63\": rpc error: code = NotFound desc = could not find container \"65653abe18fc0a6a7f40e2f85fafd80b5c9ac25b277b21ddc74f8fab2e346a63\": container with ID starting with 65653abe18fc0a6a7f40e2f85fafd80b5c9ac25b277b21ddc74f8fab2e346a63 not found: ID does not exist" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.574324 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjmd8\" (UniqueName: \"kubernetes.io/projected/b590a0dc-d944-413c-be2e-ea7933d9e9e4-kube-api-access-gjmd8\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.574391 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b590a0dc-d944-413c-be2e-ea7933d9e9e4-run-httpd\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.574459 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b590a0dc-d944-413c-be2e-ea7933d9e9e4-log-httpd\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.574787 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-config-data\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.574879 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.574965 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.575258 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-scripts\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.676627 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjmd8\" (UniqueName: \"kubernetes.io/projected/b590a0dc-d944-413c-be2e-ea7933d9e9e4-kube-api-access-gjmd8\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.676698 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b590a0dc-d944-413c-be2e-ea7933d9e9e4-run-httpd\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.676746 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b590a0dc-d944-413c-be2e-ea7933d9e9e4-log-httpd\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.676779 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-config-data\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.676802 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.676823 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.676886 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-scripts\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.677165 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b590a0dc-d944-413c-be2e-ea7933d9e9e4-log-httpd\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.677219 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b590a0dc-d944-413c-be2e-ea7933d9e9e4-run-httpd\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.684076 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-config-data\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.692718 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.693416 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjmd8\" (UniqueName: \"kubernetes.io/projected/b590a0dc-d944-413c-be2e-ea7933d9e9e4-kube-api-access-gjmd8\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.693457 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.699648 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-scripts\") pod \"ceilometer-0\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.775981 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:28:33 crc kubenswrapper[4904]: I0214 11:28:33.877999 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="562e6e60-d241-4e05-98da-80b3c449919e" path="/var/lib/kubelet/pods/562e6e60-d241-4e05-98da-80b3c449919e/volumes" Feb 14 11:28:34 crc kubenswrapper[4904]: I0214 11:28:34.274677 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:34 crc kubenswrapper[4904]: I0214 11:28:34.354474 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b590a0dc-d944-413c-be2e-ea7933d9e9e4","Type":"ContainerStarted","Data":"33f7b1879f0b681a4d0646cd67fc92ab70b0fae820c668d3d2bb5dde7ca58f96"} Feb 14 11:28:34 crc kubenswrapper[4904]: I0214 11:28:34.408744 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:38 crc kubenswrapper[4904]: I0214 11:28:38.781937 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6bc9d549b8-hdn6n" podUID="cfb252c0-ffb2-4d87-9c4e-b56cc2070d71" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 14 11:28:38 crc kubenswrapper[4904]: I0214 11:28:38.951483 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-687657b496-tz596" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 14 11:28:41 crc kubenswrapper[4904]: I0214 11:28:41.437262 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b590a0dc-d944-413c-be2e-ea7933d9e9e4","Type":"ContainerStarted","Data":"3768c88de79f10e6f25623269a486e6bdc1d014e2e9057fc89fd5446fef3e8ac"} Feb 14 11:28:41 crc kubenswrapper[4904]: I0214 11:28:41.437733 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b590a0dc-d944-413c-be2e-ea7933d9e9e4","Type":"ContainerStarted","Data":"50638b6242c7853e0f6bca5a74f696a49430021d28c350795c3e2ef5f986a091"} Feb 14 11:28:41 crc kubenswrapper[4904]: I0214 11:28:41.438967 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xdnmn" event={"ID":"6971dbc2-c11e-44d1-b63a-e0487f339183","Type":"ContainerStarted","Data":"849751e6d1a258192b1d4909642f031efd3aabe21275e10c111c340def2869e0"} Feb 14 11:28:41 crc kubenswrapper[4904]: I0214 11:28:41.456112 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-xdnmn" podStartSLOduration=2.252502177 podStartE2EDuration="12.456095027s" podCreationTimestamp="2026-02-14 11:28:29 +0000 UTC" firstStartedPulling="2026-02-14 11:28:30.443989143 +0000 UTC m=+1101.256753804" lastFinishedPulling="2026-02-14 11:28:40.647581993 +0000 UTC m=+1111.460346654" observedRunningTime="2026-02-14 11:28:41.45405399 +0000 UTC m=+1112.266818651" watchObservedRunningTime="2026-02-14 11:28:41.456095027 +0000 UTC m=+1112.268859688" Feb 14 11:28:42 crc kubenswrapper[4904]: I0214 11:28:42.450465 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b590a0dc-d944-413c-be2e-ea7933d9e9e4","Type":"ContainerStarted","Data":"b91611bb9dd7fa92d5828db01ef7d9cd8c52c856b2261816e48f3a3fc7beafb5"} Feb 14 11:28:43 crc kubenswrapper[4904]: I0214 11:28:43.463139 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b590a0dc-d944-413c-be2e-ea7933d9e9e4","Type":"ContainerStarted","Data":"9a2f9587f1979ffd4ae82561cf6db276a1697a1301b62c00fe16374d977ec528"} Feb 14 11:28:43 crc kubenswrapper[4904]: I0214 11:28:43.463380 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerName="proxy-httpd" containerID="cri-o://9a2f9587f1979ffd4ae82561cf6db276a1697a1301b62c00fe16374d977ec528" gracePeriod=30 Feb 14 11:28:43 crc kubenswrapper[4904]: I0214 11:28:43.463484 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerName="sg-core" containerID="cri-o://b91611bb9dd7fa92d5828db01ef7d9cd8c52c856b2261816e48f3a3fc7beafb5" gracePeriod=30 Feb 14 11:28:43 crc kubenswrapper[4904]: I0214 11:28:43.463596 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 14 11:28:43 crc kubenswrapper[4904]: I0214 11:28:43.463699 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerName="ceilometer-central-agent" containerID="cri-o://50638b6242c7853e0f6bca5a74f696a49430021d28c350795c3e2ef5f986a091" gracePeriod=30 Feb 14 11:28:43 crc kubenswrapper[4904]: I0214 11:28:43.463744 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerName="ceilometer-notification-agent" containerID="cri-o://3768c88de79f10e6f25623269a486e6bdc1d014e2e9057fc89fd5446fef3e8ac" gracePeriod=30 Feb 14 11:28:43 crc kubenswrapper[4904]: I0214 11:28:43.494990 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.908189632 podStartE2EDuration="10.494972322s" podCreationTimestamp="2026-02-14 11:28:33 +0000 UTC" firstStartedPulling="2026-02-14 11:28:34.285891069 +0000 UTC m=+1105.098655720" lastFinishedPulling="2026-02-14 11:28:42.872673749 +0000 UTC m=+1113.685438410" observedRunningTime="2026-02-14 11:28:43.486446318 +0000 UTC m=+1114.299210989" watchObservedRunningTime="2026-02-14 11:28:43.494972322 +0000 UTC m=+1114.307736983" Feb 14 11:28:44 crc kubenswrapper[4904]: I0214 11:28:44.482077 4904 generic.go:334] "Generic (PLEG): container finished" podID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerID="9a2f9587f1979ffd4ae82561cf6db276a1697a1301b62c00fe16374d977ec528" exitCode=0 Feb 14 11:28:44 crc kubenswrapper[4904]: I0214 11:28:44.482848 4904 generic.go:334] "Generic (PLEG): container finished" podID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerID="b91611bb9dd7fa92d5828db01ef7d9cd8c52c856b2261816e48f3a3fc7beafb5" exitCode=2 Feb 14 11:28:44 crc kubenswrapper[4904]: I0214 11:28:44.482930 4904 generic.go:334] "Generic (PLEG): container finished" podID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerID="3768c88de79f10e6f25623269a486e6bdc1d014e2e9057fc89fd5446fef3e8ac" exitCode=0 Feb 14 11:28:44 crc kubenswrapper[4904]: I0214 11:28:44.482438 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b590a0dc-d944-413c-be2e-ea7933d9e9e4","Type":"ContainerDied","Data":"9a2f9587f1979ffd4ae82561cf6db276a1697a1301b62c00fe16374d977ec528"} Feb 14 11:28:44 crc kubenswrapper[4904]: I0214 11:28:44.483076 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b590a0dc-d944-413c-be2e-ea7933d9e9e4","Type":"ContainerDied","Data":"b91611bb9dd7fa92d5828db01ef7d9cd8c52c856b2261816e48f3a3fc7beafb5"} Feb 14 11:28:44 crc kubenswrapper[4904]: I0214 11:28:44.483157 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b590a0dc-d944-413c-be2e-ea7933d9e9e4","Type":"ContainerDied","Data":"3768c88de79f10e6f25623269a486e6bdc1d014e2e9057fc89fd5446fef3e8ac"} Feb 14 11:28:46 crc kubenswrapper[4904]: I0214 11:28:46.382712 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:28:46 crc kubenswrapper[4904]: I0214 11:28:46.383047 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.517341 4904 generic.go:334] "Generic (PLEG): container finished" podID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerID="50638b6242c7853e0f6bca5a74f696a49430021d28c350795c3e2ef5f986a091" exitCode=0 Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.517914 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b590a0dc-d944-413c-be2e-ea7933d9e9e4","Type":"ContainerDied","Data":"50638b6242c7853e0f6bca5a74f696a49430021d28c350795c3e2ef5f986a091"} Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.518062 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b590a0dc-d944-413c-be2e-ea7933d9e9e4","Type":"ContainerDied","Data":"33f7b1879f0b681a4d0646cd67fc92ab70b0fae820c668d3d2bb5dde7ca58f96"} Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.518079 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33f7b1879f0b681a4d0646cd67fc92ab70b0fae820c668d3d2bb5dde7ca58f96" Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.525870 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.657236 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-scripts\") pod \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.657282 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-sg-core-conf-yaml\") pod \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.657389 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-config-data\") pod \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.657425 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b590a0dc-d944-413c-be2e-ea7933d9e9e4-log-httpd\") pod \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.657457 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjmd8\" (UniqueName: \"kubernetes.io/projected/b590a0dc-d944-413c-be2e-ea7933d9e9e4-kube-api-access-gjmd8\") pod \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.657516 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b590a0dc-d944-413c-be2e-ea7933d9e9e4-run-httpd\") pod \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.657561 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-combined-ca-bundle\") pod \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\" (UID: \"b590a0dc-d944-413c-be2e-ea7933d9e9e4\") " Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.657898 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b590a0dc-d944-413c-be2e-ea7933d9e9e4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b590a0dc-d944-413c-be2e-ea7933d9e9e4" (UID: "b590a0dc-d944-413c-be2e-ea7933d9e9e4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.658001 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b590a0dc-d944-413c-be2e-ea7933d9e9e4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b590a0dc-d944-413c-be2e-ea7933d9e9e4" (UID: "b590a0dc-d944-413c-be2e-ea7933d9e9e4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.658185 4904 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b590a0dc-d944-413c-be2e-ea7933d9e9e4-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.658198 4904 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b590a0dc-d944-413c-be2e-ea7933d9e9e4-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.663196 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-scripts" (OuterVolumeSpecName: "scripts") pod "b590a0dc-d944-413c-be2e-ea7933d9e9e4" (UID: "b590a0dc-d944-413c-be2e-ea7933d9e9e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.666849 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b590a0dc-d944-413c-be2e-ea7933d9e9e4-kube-api-access-gjmd8" (OuterVolumeSpecName: "kube-api-access-gjmd8") pod "b590a0dc-d944-413c-be2e-ea7933d9e9e4" (UID: "b590a0dc-d944-413c-be2e-ea7933d9e9e4"). InnerVolumeSpecName "kube-api-access-gjmd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.719949 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b590a0dc-d944-413c-be2e-ea7933d9e9e4" (UID: "b590a0dc-d944-413c-be2e-ea7933d9e9e4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.753602 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b590a0dc-d944-413c-be2e-ea7933d9e9e4" (UID: "b590a0dc-d944-413c-be2e-ea7933d9e9e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.759857 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjmd8\" (UniqueName: \"kubernetes.io/projected/b590a0dc-d944-413c-be2e-ea7933d9e9e4-kube-api-access-gjmd8\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.759881 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.759891 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.759901 4904 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.800060 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-config-data" (OuterVolumeSpecName: "config-data") pod "b590a0dc-d944-413c-be2e-ea7933d9e9e4" (UID: "b590a0dc-d944-413c-be2e-ea7933d9e9e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:48 crc kubenswrapper[4904]: I0214 11:28:48.862289 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b590a0dc-d944-413c-be2e-ea7933d9e9e4-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.525398 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.554652 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.564515 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.598935 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:49 crc kubenswrapper[4904]: E0214 11:28:49.599282 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerName="proxy-httpd" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.599297 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerName="proxy-httpd" Feb 14 11:28:49 crc kubenswrapper[4904]: E0214 11:28:49.599308 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerName="sg-core" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.599315 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerName="sg-core" Feb 14 11:28:49 crc kubenswrapper[4904]: E0214 11:28:49.599324 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerName="ceilometer-central-agent" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.599330 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerName="ceilometer-central-agent" Feb 14 11:28:49 crc kubenswrapper[4904]: E0214 11:28:49.599350 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerName="ceilometer-notification-agent" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.599355 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerName="ceilometer-notification-agent" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.599515 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerName="ceilometer-central-agent" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.599531 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerName="sg-core" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.599545 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerName="proxy-httpd" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.599559 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" containerName="ceilometer-notification-agent" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.602332 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.605564 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.605853 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.659891 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.676737 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-config-data\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.677159 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37150fd8-4835-4de8-9867-7da3176031a2-log-httpd\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.677208 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.677272 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wpqt\" (UniqueName: \"kubernetes.io/projected/37150fd8-4835-4de8-9867-7da3176031a2-kube-api-access-6wpqt\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.677299 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.677327 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-scripts\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.677377 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37150fd8-4835-4de8-9867-7da3176031a2-run-httpd\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.778441 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wpqt\" (UniqueName: \"kubernetes.io/projected/37150fd8-4835-4de8-9867-7da3176031a2-kube-api-access-6wpqt\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.778489 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.778515 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-scripts\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.778555 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37150fd8-4835-4de8-9867-7da3176031a2-run-httpd\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.778581 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-config-data\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.778632 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37150fd8-4835-4de8-9867-7da3176031a2-log-httpd\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.778689 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.779219 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37150fd8-4835-4de8-9867-7da3176031a2-log-httpd\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.779539 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37150fd8-4835-4de8-9867-7da3176031a2-run-httpd\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.785969 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-config-data\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.786536 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-scripts\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.789658 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.794573 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.809751 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wpqt\" (UniqueName: \"kubernetes.io/projected/37150fd8-4835-4de8-9867-7da3176031a2-kube-api-access-6wpqt\") pod \"ceilometer-0\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " pod="openstack/ceilometer-0" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.861954 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b590a0dc-d944-413c-be2e-ea7933d9e9e4" path="/var/lib/kubelet/pods/b590a0dc-d944-413c-be2e-ea7933d9e9e4/volumes" Feb 14 11:28:49 crc kubenswrapper[4904]: I0214 11:28:49.924354 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:28:50 crc kubenswrapper[4904]: I0214 11:28:50.443795 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:50 crc kubenswrapper[4904]: W0214 11:28:50.446800 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37150fd8_4835_4de8_9867_7da3176031a2.slice/crio-815be35d911cac415fde04c684f24d6a5d7d402676b78a4ff8ef6954a8914326 WatchSource:0}: Error finding container 815be35d911cac415fde04c684f24d6a5d7d402676b78a4ff8ef6954a8914326: Status 404 returned error can't find the container with id 815be35d911cac415fde04c684f24d6a5d7d402676b78a4ff8ef6954a8914326 Feb 14 11:28:50 crc kubenswrapper[4904]: I0214 11:28:50.540471 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37150fd8-4835-4de8-9867-7da3176031a2","Type":"ContainerStarted","Data":"815be35d911cac415fde04c684f24d6a5d7d402676b78a4ff8ef6954a8914326"} Feb 14 11:28:51 crc kubenswrapper[4904]: I0214 11:28:51.352469 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:51 crc kubenswrapper[4904]: I0214 11:28:51.554549 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37150fd8-4835-4de8-9867-7da3176031a2","Type":"ContainerStarted","Data":"1e7962c18f7ab4583e307f3e8278733ab88a16e2142e9a15e154e542eabbb6fb"} Feb 14 11:28:51 crc kubenswrapper[4904]: I0214 11:28:51.605526 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-687657b496-tz596" Feb 14 11:28:51 crc kubenswrapper[4904]: I0214 11:28:51.636769 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:28:52 crc kubenswrapper[4904]: I0214 11:28:52.566157 4904 generic.go:334] "Generic (PLEG): container finished" podID="6971dbc2-c11e-44d1-b63a-e0487f339183" containerID="849751e6d1a258192b1d4909642f031efd3aabe21275e10c111c340def2869e0" exitCode=0 Feb 14 11:28:52 crc kubenswrapper[4904]: I0214 11:28:52.566325 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xdnmn" event={"ID":"6971dbc2-c11e-44d1-b63a-e0487f339183","Type":"ContainerDied","Data":"849751e6d1a258192b1d4909642f031efd3aabe21275e10c111c340def2869e0"} Feb 14 11:28:52 crc kubenswrapper[4904]: I0214 11:28:52.569136 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37150fd8-4835-4de8-9867-7da3176031a2","Type":"ContainerStarted","Data":"679f00ecb817e2d7295f9b2fcc4b6362ce0e4534ed4d472e397427ca49010109"} Feb 14 11:28:52 crc kubenswrapper[4904]: I0214 11:28:52.569167 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37150fd8-4835-4de8-9867-7da3176031a2","Type":"ContainerStarted","Data":"977a2cf5a4732f409143a9a688889694570c2298b631cb27a26e3e0064d8d77c"} Feb 14 11:28:53 crc kubenswrapper[4904]: I0214 11:28:53.532189 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6bc9d549b8-hdn6n" Feb 14 11:28:53 crc kubenswrapper[4904]: I0214 11:28:53.623718 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-687657b496-tz596"] Feb 14 11:28:53 crc kubenswrapper[4904]: I0214 11:28:53.624009 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-687657b496-tz596" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon-log" containerID="cri-o://582bc45e7c749f481aab958e03a53b45ff101180b75c9d8a70ed2568b4a2948b" gracePeriod=30 Feb 14 11:28:53 crc kubenswrapper[4904]: I0214 11:28:53.624932 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-687657b496-tz596" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon" containerID="cri-o://a6c1837a2344350a20faa0c4c53cf29edd5c3937922bb9a13dd8aa63719d0470" gracePeriod=30 Feb 14 11:28:53 crc kubenswrapper[4904]: I0214 11:28:53.642986 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-687657b496-tz596" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.470908 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xdnmn" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.595871 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xdnmn" event={"ID":"6971dbc2-c11e-44d1-b63a-e0487f339183","Type":"ContainerDied","Data":"9bcb8f260ae0e4278935c16f871cb009260ad6ba0d37d36f7c0473ae5984437d"} Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.596289 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bcb8f260ae0e4278935c16f871cb009260ad6ba0d37d36f7c0473ae5984437d" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.595901 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xdnmn" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.599203 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37150fd8-4835-4de8-9867-7da3176031a2","Type":"ContainerStarted","Data":"325682c5f98c84eb574e1c22311bca7854595c6fa809b281e2930d884d8b4674"} Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.599332 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37150fd8-4835-4de8-9867-7da3176031a2" containerName="ceilometer-central-agent" containerID="cri-o://1e7962c18f7ab4583e307f3e8278733ab88a16e2142e9a15e154e542eabbb6fb" gracePeriod=30 Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.599427 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.599764 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37150fd8-4835-4de8-9867-7da3176031a2" containerName="proxy-httpd" containerID="cri-o://325682c5f98c84eb574e1c22311bca7854595c6fa809b281e2930d884d8b4674" gracePeriod=30 Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.599808 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37150fd8-4835-4de8-9867-7da3176031a2" containerName="sg-core" containerID="cri-o://679f00ecb817e2d7295f9b2fcc4b6362ce0e4534ed4d472e397427ca49010109" gracePeriod=30 Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.599879 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37150fd8-4835-4de8-9867-7da3176031a2" containerName="ceilometer-notification-agent" containerID="cri-o://977a2cf5a4732f409143a9a688889694570c2298b631cb27a26e3e0064d8d77c" gracePeriod=30 Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.619861 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6971dbc2-c11e-44d1-b63a-e0487f339183-scripts\") pod \"6971dbc2-c11e-44d1-b63a-e0487f339183\" (UID: \"6971dbc2-c11e-44d1-b63a-e0487f339183\") " Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.619944 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmjfp\" (UniqueName: \"kubernetes.io/projected/6971dbc2-c11e-44d1-b63a-e0487f339183-kube-api-access-nmjfp\") pod \"6971dbc2-c11e-44d1-b63a-e0487f339183\" (UID: \"6971dbc2-c11e-44d1-b63a-e0487f339183\") " Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.620570 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6971dbc2-c11e-44d1-b63a-e0487f339183-config-data\") pod \"6971dbc2-c11e-44d1-b63a-e0487f339183\" (UID: \"6971dbc2-c11e-44d1-b63a-e0487f339183\") " Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.620627 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6971dbc2-c11e-44d1-b63a-e0487f339183-combined-ca-bundle\") pod \"6971dbc2-c11e-44d1-b63a-e0487f339183\" (UID: \"6971dbc2-c11e-44d1-b63a-e0487f339183\") " Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.634222 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6971dbc2-c11e-44d1-b63a-e0487f339183-kube-api-access-nmjfp" (OuterVolumeSpecName: "kube-api-access-nmjfp") pod "6971dbc2-c11e-44d1-b63a-e0487f339183" (UID: "6971dbc2-c11e-44d1-b63a-e0487f339183"). InnerVolumeSpecName "kube-api-access-nmjfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.646053 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.764781195 podStartE2EDuration="5.646034695s" podCreationTimestamp="2026-02-14 11:28:49 +0000 UTC" firstStartedPulling="2026-02-14 11:28:50.449415256 +0000 UTC m=+1121.262179927" lastFinishedPulling="2026-02-14 11:28:53.330668766 +0000 UTC m=+1124.143433427" observedRunningTime="2026-02-14 11:28:54.634658243 +0000 UTC m=+1125.447422894" watchObservedRunningTime="2026-02-14 11:28:54.646034695 +0000 UTC m=+1125.458799356" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.646397 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6971dbc2-c11e-44d1-b63a-e0487f339183-scripts" (OuterVolumeSpecName: "scripts") pod "6971dbc2-c11e-44d1-b63a-e0487f339183" (UID: "6971dbc2-c11e-44d1-b63a-e0487f339183"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.686980 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6971dbc2-c11e-44d1-b63a-e0487f339183-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6971dbc2-c11e-44d1-b63a-e0487f339183" (UID: "6971dbc2-c11e-44d1-b63a-e0487f339183"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.712341 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6971dbc2-c11e-44d1-b63a-e0487f339183-config-data" (OuterVolumeSpecName: "config-data") pod "6971dbc2-c11e-44d1-b63a-e0487f339183" (UID: "6971dbc2-c11e-44d1-b63a-e0487f339183"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.722089 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6971dbc2-c11e-44d1-b63a-e0487f339183-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.722116 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6971dbc2-c11e-44d1-b63a-e0487f339183-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.722125 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmjfp\" (UniqueName: \"kubernetes.io/projected/6971dbc2-c11e-44d1-b63a-e0487f339183-kube-api-access-nmjfp\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.722136 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6971dbc2-c11e-44d1-b63a-e0487f339183-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.745157 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 14 11:28:54 crc kubenswrapper[4904]: E0214 11:28:54.745769 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6971dbc2-c11e-44d1-b63a-e0487f339183" containerName="nova-cell0-conductor-db-sync" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.745848 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="6971dbc2-c11e-44d1-b63a-e0487f339183" containerName="nova-cell0-conductor-db-sync" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.746231 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="6971dbc2-c11e-44d1-b63a-e0487f339183" containerName="nova-cell0-conductor-db-sync" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.746850 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.778031 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.925608 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1539e687-1fb9-4f6b-ba62-204f09bb691f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1539e687-1fb9-4f6b-ba62-204f09bb691f\") " pod="openstack/nova-cell0-conductor-0" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.925682 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz6kn\" (UniqueName: \"kubernetes.io/projected/1539e687-1fb9-4f6b-ba62-204f09bb691f-kube-api-access-pz6kn\") pod \"nova-cell0-conductor-0\" (UID: \"1539e687-1fb9-4f6b-ba62-204f09bb691f\") " pod="openstack/nova-cell0-conductor-0" Feb 14 11:28:54 crc kubenswrapper[4904]: I0214 11:28:54.925722 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1539e687-1fb9-4f6b-ba62-204f09bb691f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1539e687-1fb9-4f6b-ba62-204f09bb691f\") " pod="openstack/nova-cell0-conductor-0" Feb 14 11:28:55 crc kubenswrapper[4904]: I0214 11:28:55.027880 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1539e687-1fb9-4f6b-ba62-204f09bb691f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1539e687-1fb9-4f6b-ba62-204f09bb691f\") " pod="openstack/nova-cell0-conductor-0" Feb 14 11:28:55 crc kubenswrapper[4904]: I0214 11:28:55.027964 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz6kn\" (UniqueName: \"kubernetes.io/projected/1539e687-1fb9-4f6b-ba62-204f09bb691f-kube-api-access-pz6kn\") pod \"nova-cell0-conductor-0\" (UID: \"1539e687-1fb9-4f6b-ba62-204f09bb691f\") " pod="openstack/nova-cell0-conductor-0" Feb 14 11:28:55 crc kubenswrapper[4904]: I0214 11:28:55.027995 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1539e687-1fb9-4f6b-ba62-204f09bb691f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1539e687-1fb9-4f6b-ba62-204f09bb691f\") " pod="openstack/nova-cell0-conductor-0" Feb 14 11:28:55 crc kubenswrapper[4904]: I0214 11:28:55.073148 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1539e687-1fb9-4f6b-ba62-204f09bb691f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1539e687-1fb9-4f6b-ba62-204f09bb691f\") " pod="openstack/nova-cell0-conductor-0" Feb 14 11:28:55 crc kubenswrapper[4904]: I0214 11:28:55.074107 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1539e687-1fb9-4f6b-ba62-204f09bb691f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1539e687-1fb9-4f6b-ba62-204f09bb691f\") " pod="openstack/nova-cell0-conductor-0" Feb 14 11:28:55 crc kubenswrapper[4904]: I0214 11:28:55.092584 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz6kn\" (UniqueName: \"kubernetes.io/projected/1539e687-1fb9-4f6b-ba62-204f09bb691f-kube-api-access-pz6kn\") pod \"nova-cell0-conductor-0\" (UID: \"1539e687-1fb9-4f6b-ba62-204f09bb691f\") " pod="openstack/nova-cell0-conductor-0" Feb 14 11:28:55 crc kubenswrapper[4904]: I0214 11:28:55.097502 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 14 11:28:55 crc kubenswrapper[4904]: I0214 11:28:55.595519 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 14 11:28:55 crc kubenswrapper[4904]: I0214 11:28:55.655614 4904 generic.go:334] "Generic (PLEG): container finished" podID="37150fd8-4835-4de8-9867-7da3176031a2" containerID="325682c5f98c84eb574e1c22311bca7854595c6fa809b281e2930d884d8b4674" exitCode=0 Feb 14 11:28:55 crc kubenswrapper[4904]: I0214 11:28:55.655648 4904 generic.go:334] "Generic (PLEG): container finished" podID="37150fd8-4835-4de8-9867-7da3176031a2" containerID="679f00ecb817e2d7295f9b2fcc4b6362ce0e4534ed4d472e397427ca49010109" exitCode=2 Feb 14 11:28:55 crc kubenswrapper[4904]: I0214 11:28:55.655658 4904 generic.go:334] "Generic (PLEG): container finished" podID="37150fd8-4835-4de8-9867-7da3176031a2" containerID="977a2cf5a4732f409143a9a688889694570c2298b631cb27a26e3e0064d8d77c" exitCode=0 Feb 14 11:28:55 crc kubenswrapper[4904]: I0214 11:28:55.655705 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37150fd8-4835-4de8-9867-7da3176031a2","Type":"ContainerDied","Data":"325682c5f98c84eb574e1c22311bca7854595c6fa809b281e2930d884d8b4674"} Feb 14 11:28:55 crc kubenswrapper[4904]: I0214 11:28:55.656443 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37150fd8-4835-4de8-9867-7da3176031a2","Type":"ContainerDied","Data":"679f00ecb817e2d7295f9b2fcc4b6362ce0e4534ed4d472e397427ca49010109"} Feb 14 11:28:55 crc kubenswrapper[4904]: I0214 11:28:55.656467 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37150fd8-4835-4de8-9867-7da3176031a2","Type":"ContainerDied","Data":"977a2cf5a4732f409143a9a688889694570c2298b631cb27a26e3e0064d8d77c"} Feb 14 11:28:55 crc kubenswrapper[4904]: I0214 11:28:55.664890 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1539e687-1fb9-4f6b-ba62-204f09bb691f","Type":"ContainerStarted","Data":"8f3356762558311cae32649510b49e7c80a3bc455f9e93744c265583a2bd45cd"} Feb 14 11:28:55 crc kubenswrapper[4904]: I0214 11:28:55.774895 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-687657b496-tz596" Feb 14 11:28:56 crc kubenswrapper[4904]: I0214 11:28:56.675406 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1539e687-1fb9-4f6b-ba62-204f09bb691f","Type":"ContainerStarted","Data":"5f0304219bc323401e1915fac605ccd6d6eb89ca0f1ef9897bcec13e7d6ee844"} Feb 14 11:28:56 crc kubenswrapper[4904]: I0214 11:28:56.675802 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 14 11:28:56 crc kubenswrapper[4904]: I0214 11:28:56.695877 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.695853222 podStartE2EDuration="2.695853222s" podCreationTimestamp="2026-02-14 11:28:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:28:56.690547546 +0000 UTC m=+1127.503312207" watchObservedRunningTime="2026-02-14 11:28:56.695853222 +0000 UTC m=+1127.508617893" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.562351 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.693620 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37150fd8-4835-4de8-9867-7da3176031a2-run-httpd\") pod \"37150fd8-4835-4de8-9867-7da3176031a2\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.693666 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-combined-ca-bundle\") pod \"37150fd8-4835-4de8-9867-7da3176031a2\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.693804 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-scripts\") pod \"37150fd8-4835-4de8-9867-7da3176031a2\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.693851 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wpqt\" (UniqueName: \"kubernetes.io/projected/37150fd8-4835-4de8-9867-7da3176031a2-kube-api-access-6wpqt\") pod \"37150fd8-4835-4de8-9867-7da3176031a2\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.693877 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-config-data\") pod \"37150fd8-4835-4de8-9867-7da3176031a2\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.693907 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-sg-core-conf-yaml\") pod \"37150fd8-4835-4de8-9867-7da3176031a2\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.693976 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37150fd8-4835-4de8-9867-7da3176031a2-log-httpd\") pod \"37150fd8-4835-4de8-9867-7da3176031a2\" (UID: \"37150fd8-4835-4de8-9867-7da3176031a2\") " Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.694141 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37150fd8-4835-4de8-9867-7da3176031a2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "37150fd8-4835-4de8-9867-7da3176031a2" (UID: "37150fd8-4835-4de8-9867-7da3176031a2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.694477 4904 generic.go:334] "Generic (PLEG): container finished" podID="37150fd8-4835-4de8-9867-7da3176031a2" containerID="1e7962c18f7ab4583e307f3e8278733ab88a16e2142e9a15e154e542eabbb6fb" exitCode=0 Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.694511 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37150fd8-4835-4de8-9867-7da3176031a2","Type":"ContainerDied","Data":"1e7962c18f7ab4583e307f3e8278733ab88a16e2142e9a15e154e542eabbb6fb"} Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.694536 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37150fd8-4835-4de8-9867-7da3176031a2","Type":"ContainerDied","Data":"815be35d911cac415fde04c684f24d6a5d7d402676b78a4ff8ef6954a8914326"} Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.694552 4904 scope.go:117] "RemoveContainer" containerID="325682c5f98c84eb574e1c22311bca7854595c6fa809b281e2930d884d8b4674" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.694659 4904 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37150fd8-4835-4de8-9867-7da3176031a2-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.694670 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.694847 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37150fd8-4835-4de8-9867-7da3176031a2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "37150fd8-4835-4de8-9867-7da3176031a2" (UID: "37150fd8-4835-4de8-9867-7da3176031a2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.699723 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37150fd8-4835-4de8-9867-7da3176031a2-kube-api-access-6wpqt" (OuterVolumeSpecName: "kube-api-access-6wpqt") pod "37150fd8-4835-4de8-9867-7da3176031a2" (UID: "37150fd8-4835-4de8-9867-7da3176031a2"). InnerVolumeSpecName "kube-api-access-6wpqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.701502 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-scripts" (OuterVolumeSpecName: "scripts") pod "37150fd8-4835-4de8-9867-7da3176031a2" (UID: "37150fd8-4835-4de8-9867-7da3176031a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.719600 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "37150fd8-4835-4de8-9867-7da3176031a2" (UID: "37150fd8-4835-4de8-9867-7da3176031a2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.771822 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37150fd8-4835-4de8-9867-7da3176031a2" (UID: "37150fd8-4835-4de8-9867-7da3176031a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.794302 4904 scope.go:117] "RemoveContainer" containerID="679f00ecb817e2d7295f9b2fcc4b6362ce0e4534ed4d472e397427ca49010109" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.796281 4904 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37150fd8-4835-4de8-9867-7da3176031a2-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.796614 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.796683 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.796745 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wpqt\" (UniqueName: \"kubernetes.io/projected/37150fd8-4835-4de8-9867-7da3176031a2-kube-api-access-6wpqt\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.796802 4904 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.811406 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-config-data" (OuterVolumeSpecName: "config-data") pod "37150fd8-4835-4de8-9867-7da3176031a2" (UID: "37150fd8-4835-4de8-9867-7da3176031a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.816306 4904 scope.go:117] "RemoveContainer" containerID="977a2cf5a4732f409143a9a688889694570c2298b631cb27a26e3e0064d8d77c" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.859396 4904 scope.go:117] "RemoveContainer" containerID="1e7962c18f7ab4583e307f3e8278733ab88a16e2142e9a15e154e542eabbb6fb" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.884040 4904 scope.go:117] "RemoveContainer" containerID="325682c5f98c84eb574e1c22311bca7854595c6fa809b281e2930d884d8b4674" Feb 14 11:28:58 crc kubenswrapper[4904]: E0214 11:28:58.884501 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"325682c5f98c84eb574e1c22311bca7854595c6fa809b281e2930d884d8b4674\": container with ID starting with 325682c5f98c84eb574e1c22311bca7854595c6fa809b281e2930d884d8b4674 not found: ID does not exist" containerID="325682c5f98c84eb574e1c22311bca7854595c6fa809b281e2930d884d8b4674" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.884532 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"325682c5f98c84eb574e1c22311bca7854595c6fa809b281e2930d884d8b4674"} err="failed to get container status \"325682c5f98c84eb574e1c22311bca7854595c6fa809b281e2930d884d8b4674\": rpc error: code = NotFound desc = could not find container \"325682c5f98c84eb574e1c22311bca7854595c6fa809b281e2930d884d8b4674\": container with ID starting with 325682c5f98c84eb574e1c22311bca7854595c6fa809b281e2930d884d8b4674 not found: ID does not exist" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.884552 4904 scope.go:117] "RemoveContainer" containerID="679f00ecb817e2d7295f9b2fcc4b6362ce0e4534ed4d472e397427ca49010109" Feb 14 11:28:58 crc kubenswrapper[4904]: E0214 11:28:58.885202 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"679f00ecb817e2d7295f9b2fcc4b6362ce0e4534ed4d472e397427ca49010109\": container with ID starting with 679f00ecb817e2d7295f9b2fcc4b6362ce0e4534ed4d472e397427ca49010109 not found: ID does not exist" containerID="679f00ecb817e2d7295f9b2fcc4b6362ce0e4534ed4d472e397427ca49010109" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.885225 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"679f00ecb817e2d7295f9b2fcc4b6362ce0e4534ed4d472e397427ca49010109"} err="failed to get container status \"679f00ecb817e2d7295f9b2fcc4b6362ce0e4534ed4d472e397427ca49010109\": rpc error: code = NotFound desc = could not find container \"679f00ecb817e2d7295f9b2fcc4b6362ce0e4534ed4d472e397427ca49010109\": container with ID starting with 679f00ecb817e2d7295f9b2fcc4b6362ce0e4534ed4d472e397427ca49010109 not found: ID does not exist" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.885241 4904 scope.go:117] "RemoveContainer" containerID="977a2cf5a4732f409143a9a688889694570c2298b631cb27a26e3e0064d8d77c" Feb 14 11:28:58 crc kubenswrapper[4904]: E0214 11:28:58.885474 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"977a2cf5a4732f409143a9a688889694570c2298b631cb27a26e3e0064d8d77c\": container with ID starting with 977a2cf5a4732f409143a9a688889694570c2298b631cb27a26e3e0064d8d77c not found: ID does not exist" containerID="977a2cf5a4732f409143a9a688889694570c2298b631cb27a26e3e0064d8d77c" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.885494 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"977a2cf5a4732f409143a9a688889694570c2298b631cb27a26e3e0064d8d77c"} err="failed to get container status \"977a2cf5a4732f409143a9a688889694570c2298b631cb27a26e3e0064d8d77c\": rpc error: code = NotFound desc = could not find container \"977a2cf5a4732f409143a9a688889694570c2298b631cb27a26e3e0064d8d77c\": container with ID starting with 977a2cf5a4732f409143a9a688889694570c2298b631cb27a26e3e0064d8d77c not found: ID does not exist" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.885507 4904 scope.go:117] "RemoveContainer" containerID="1e7962c18f7ab4583e307f3e8278733ab88a16e2142e9a15e154e542eabbb6fb" Feb 14 11:28:58 crc kubenswrapper[4904]: E0214 11:28:58.885807 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e7962c18f7ab4583e307f3e8278733ab88a16e2142e9a15e154e542eabbb6fb\": container with ID starting with 1e7962c18f7ab4583e307f3e8278733ab88a16e2142e9a15e154e542eabbb6fb not found: ID does not exist" containerID="1e7962c18f7ab4583e307f3e8278733ab88a16e2142e9a15e154e542eabbb6fb" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.885827 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e7962c18f7ab4583e307f3e8278733ab88a16e2142e9a15e154e542eabbb6fb"} err="failed to get container status \"1e7962c18f7ab4583e307f3e8278733ab88a16e2142e9a15e154e542eabbb6fb\": rpc error: code = NotFound desc = could not find container \"1e7962c18f7ab4583e307f3e8278733ab88a16e2142e9a15e154e542eabbb6fb\": container with ID starting with 1e7962c18f7ab4583e307f3e8278733ab88a16e2142e9a15e154e542eabbb6fb not found: ID does not exist" Feb 14 11:28:58 crc kubenswrapper[4904]: I0214 11:28:58.898398 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37150fd8-4835-4de8-9867-7da3176031a2-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.033153 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.045770 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.066550 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:59 crc kubenswrapper[4904]: E0214 11:28:59.066879 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37150fd8-4835-4de8-9867-7da3176031a2" containerName="ceilometer-central-agent" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.066896 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="37150fd8-4835-4de8-9867-7da3176031a2" containerName="ceilometer-central-agent" Feb 14 11:28:59 crc kubenswrapper[4904]: E0214 11:28:59.066907 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37150fd8-4835-4de8-9867-7da3176031a2" containerName="sg-core" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.066914 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="37150fd8-4835-4de8-9867-7da3176031a2" containerName="sg-core" Feb 14 11:28:59 crc kubenswrapper[4904]: E0214 11:28:59.066928 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37150fd8-4835-4de8-9867-7da3176031a2" containerName="proxy-httpd" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.066934 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="37150fd8-4835-4de8-9867-7da3176031a2" containerName="proxy-httpd" Feb 14 11:28:59 crc kubenswrapper[4904]: E0214 11:28:59.066950 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37150fd8-4835-4de8-9867-7da3176031a2" containerName="ceilometer-notification-agent" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.066957 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="37150fd8-4835-4de8-9867-7da3176031a2" containerName="ceilometer-notification-agent" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.067121 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="37150fd8-4835-4de8-9867-7da3176031a2" containerName="sg-core" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.067133 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="37150fd8-4835-4de8-9867-7da3176031a2" containerName="ceilometer-central-agent" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.067146 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="37150fd8-4835-4de8-9867-7da3176031a2" containerName="proxy-httpd" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.067158 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="37150fd8-4835-4de8-9867-7da3176031a2" containerName="ceilometer-notification-agent" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.070149 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.073252 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.073433 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.090915 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.142502 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-687657b496-tz596" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:49104->10.217.0.150:8443: read: connection reset by peer" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.204037 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-config-data\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.204363 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.204507 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-scripts\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.204689 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3169126-e313-4d31-8984-34e321aa0bde-log-httpd\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.204809 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shrjv\" (UniqueName: \"kubernetes.io/projected/e3169126-e313-4d31-8984-34e321aa0bde-kube-api-access-shrjv\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.204987 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.205100 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3169126-e313-4d31-8984-34e321aa0bde-run-httpd\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.306715 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-config-data\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.307306 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.307334 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-scripts\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.307710 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3169126-e313-4d31-8984-34e321aa0bde-log-httpd\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.307750 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shrjv\" (UniqueName: \"kubernetes.io/projected/e3169126-e313-4d31-8984-34e321aa0bde-kube-api-access-shrjv\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.307865 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.307890 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3169126-e313-4d31-8984-34e321aa0bde-run-httpd\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.308271 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3169126-e313-4d31-8984-34e321aa0bde-log-httpd\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.308301 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3169126-e313-4d31-8984-34e321aa0bde-run-httpd\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.311612 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.312266 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-scripts\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.313288 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.316354 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-config-data\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.326222 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shrjv\" (UniqueName: \"kubernetes.io/projected/e3169126-e313-4d31-8984-34e321aa0bde-kube-api-access-shrjv\") pod \"ceilometer-0\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.387254 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.711492 4904 generic.go:334] "Generic (PLEG): container finished" podID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerID="a6c1837a2344350a20faa0c4c53cf29edd5c3937922bb9a13dd8aa63719d0470" exitCode=0 Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.711570 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-687657b496-tz596" event={"ID":"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393","Type":"ContainerDied","Data":"a6c1837a2344350a20faa0c4c53cf29edd5c3937922bb9a13dd8aa63719d0470"} Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.711874 4904 scope.go:117] "RemoveContainer" containerID="609e50be7f47738cb94497fefee796e278d503fdc50869f2e264123ccbb09fa6" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.857119 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37150fd8-4835-4de8-9867-7da3176031a2" path="/var/lib/kubelet/pods/37150fd8-4835-4de8-9867-7da3176031a2/volumes" Feb 14 11:28:59 crc kubenswrapper[4904]: I0214 11:28:59.894825 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:28:59 crc kubenswrapper[4904]: W0214 11:28:59.905854 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3169126_e313_4d31_8984_34e321aa0bde.slice/crio-8053e19b9a6494b14671b7fe2c963ab59d4bf1dee43b6803b3c1b1bbaed05490 WatchSource:0}: Error finding container 8053e19b9a6494b14671b7fe2c963ab59d4bf1dee43b6803b3c1b1bbaed05490: Status 404 returned error can't find the container with id 8053e19b9a6494b14671b7fe2c963ab59d4bf1dee43b6803b3c1b1bbaed05490 Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.131248 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.664391 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-98lq8"] Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.665803 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-98lq8" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.669299 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.670687 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.679441 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-98lq8"] Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.726285 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3169126-e313-4d31-8984-34e321aa0bde","Type":"ContainerStarted","Data":"5a83e64acd6041c6fc9c5b408baa8ff73b6d313aa58936932d5b78aa7c9c9fba"} Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.726335 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3169126-e313-4d31-8984-34e321aa0bde","Type":"ContainerStarted","Data":"8053e19b9a6494b14671b7fe2c963ab59d4bf1dee43b6803b3c1b1bbaed05490"} Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.738822 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/806d3bba-5958-4939-90a5-ae94621e1c5a-config-data\") pod \"nova-cell0-cell-mapping-98lq8\" (UID: \"806d3bba-5958-4939-90a5-ae94621e1c5a\") " pod="openstack/nova-cell0-cell-mapping-98lq8" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.738904 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/806d3bba-5958-4939-90a5-ae94621e1c5a-scripts\") pod \"nova-cell0-cell-mapping-98lq8\" (UID: \"806d3bba-5958-4939-90a5-ae94621e1c5a\") " pod="openstack/nova-cell0-cell-mapping-98lq8" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.738968 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjmvl\" (UniqueName: \"kubernetes.io/projected/806d3bba-5958-4939-90a5-ae94621e1c5a-kube-api-access-kjmvl\") pod \"nova-cell0-cell-mapping-98lq8\" (UID: \"806d3bba-5958-4939-90a5-ae94621e1c5a\") " pod="openstack/nova-cell0-cell-mapping-98lq8" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.739094 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/806d3bba-5958-4939-90a5-ae94621e1c5a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-98lq8\" (UID: \"806d3bba-5958-4939-90a5-ae94621e1c5a\") " pod="openstack/nova-cell0-cell-mapping-98lq8" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.841149 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/806d3bba-5958-4939-90a5-ae94621e1c5a-config-data\") pod \"nova-cell0-cell-mapping-98lq8\" (UID: \"806d3bba-5958-4939-90a5-ae94621e1c5a\") " pod="openstack/nova-cell0-cell-mapping-98lq8" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.841196 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/806d3bba-5958-4939-90a5-ae94621e1c5a-scripts\") pod \"nova-cell0-cell-mapping-98lq8\" (UID: \"806d3bba-5958-4939-90a5-ae94621e1c5a\") " pod="openstack/nova-cell0-cell-mapping-98lq8" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.841252 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjmvl\" (UniqueName: \"kubernetes.io/projected/806d3bba-5958-4939-90a5-ae94621e1c5a-kube-api-access-kjmvl\") pod \"nova-cell0-cell-mapping-98lq8\" (UID: \"806d3bba-5958-4939-90a5-ae94621e1c5a\") " pod="openstack/nova-cell0-cell-mapping-98lq8" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.841310 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/806d3bba-5958-4939-90a5-ae94621e1c5a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-98lq8\" (UID: \"806d3bba-5958-4939-90a5-ae94621e1c5a\") " pod="openstack/nova-cell0-cell-mapping-98lq8" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.867989 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/806d3bba-5958-4939-90a5-ae94621e1c5a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-98lq8\" (UID: \"806d3bba-5958-4939-90a5-ae94621e1c5a\") " pod="openstack/nova-cell0-cell-mapping-98lq8" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.869331 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/806d3bba-5958-4939-90a5-ae94621e1c5a-scripts\") pod \"nova-cell0-cell-mapping-98lq8\" (UID: \"806d3bba-5958-4939-90a5-ae94621e1c5a\") " pod="openstack/nova-cell0-cell-mapping-98lq8" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.875159 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/806d3bba-5958-4939-90a5-ae94621e1c5a-config-data\") pod \"nova-cell0-cell-mapping-98lq8\" (UID: \"806d3bba-5958-4939-90a5-ae94621e1c5a\") " pod="openstack/nova-cell0-cell-mapping-98lq8" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.876791 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.878110 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.883681 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.897561 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjmvl\" (UniqueName: \"kubernetes.io/projected/806d3bba-5958-4939-90a5-ae94621e1c5a-kube-api-access-kjmvl\") pod \"nova-cell0-cell-mapping-98lq8\" (UID: \"806d3bba-5958-4939-90a5-ae94621e1c5a\") " pod="openstack/nova-cell0-cell-mapping-98lq8" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.902356 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.942437 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjk56\" (UniqueName: \"kubernetes.io/projected/c97027f1-aa8d-4457-a4f9-31812b8107d3-kube-api-access-mjk56\") pod \"nova-scheduler-0\" (UID: \"c97027f1-aa8d-4457-a4f9-31812b8107d3\") " pod="openstack/nova-scheduler-0" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.942597 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c97027f1-aa8d-4457-a4f9-31812b8107d3-config-data\") pod \"nova-scheduler-0\" (UID: \"c97027f1-aa8d-4457-a4f9-31812b8107d3\") " pod="openstack/nova-scheduler-0" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.942666 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c97027f1-aa8d-4457-a4f9-31812b8107d3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c97027f1-aa8d-4457-a4f9-31812b8107d3\") " pod="openstack/nova-scheduler-0" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.957256 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.961803 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.976700 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 14 11:29:00 crc kubenswrapper[4904]: I0214 11:29:00.988465 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-98lq8" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.045206 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c97027f1-aa8d-4457-a4f9-31812b8107d3-config-data\") pod \"nova-scheduler-0\" (UID: \"c97027f1-aa8d-4457-a4f9-31812b8107d3\") " pod="openstack/nova-scheduler-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.070243 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17d3d79f-6c72-4ad5-862e-7a1e020333ae-logs\") pod \"nova-api-0\" (UID: \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\") " pod="openstack/nova-api-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.070606 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17d3d79f-6c72-4ad5-862e-7a1e020333ae-config-data\") pod \"nova-api-0\" (UID: \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\") " pod="openstack/nova-api-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.070714 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nwlg\" (UniqueName: \"kubernetes.io/projected/17d3d79f-6c72-4ad5-862e-7a1e020333ae-kube-api-access-5nwlg\") pod \"nova-api-0\" (UID: \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\") " pod="openstack/nova-api-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.070740 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c97027f1-aa8d-4457-a4f9-31812b8107d3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c97027f1-aa8d-4457-a4f9-31812b8107d3\") " pod="openstack/nova-scheduler-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.070797 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjk56\" (UniqueName: \"kubernetes.io/projected/c97027f1-aa8d-4457-a4f9-31812b8107d3-kube-api-access-mjk56\") pod \"nova-scheduler-0\" (UID: \"c97027f1-aa8d-4457-a4f9-31812b8107d3\") " pod="openstack/nova-scheduler-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.070892 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17d3d79f-6c72-4ad5-862e-7a1e020333ae-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\") " pod="openstack/nova-api-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.049511 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c97027f1-aa8d-4457-a4f9-31812b8107d3-config-data\") pod \"nova-scheduler-0\" (UID: \"c97027f1-aa8d-4457-a4f9-31812b8107d3\") " pod="openstack/nova-scheduler-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.072226 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.085078 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c97027f1-aa8d-4457-a4f9-31812b8107d3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c97027f1-aa8d-4457-a4f9-31812b8107d3\") " pod="openstack/nova-scheduler-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.121449 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjk56\" (UniqueName: \"kubernetes.io/projected/c97027f1-aa8d-4457-a4f9-31812b8107d3-kube-api-access-mjk56\") pod \"nova-scheduler-0\" (UID: \"c97027f1-aa8d-4457-a4f9-31812b8107d3\") " pod="openstack/nova-scheduler-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.134925 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.139090 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.148855 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.172609 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17d3d79f-6c72-4ad5-862e-7a1e020333ae-config-data\") pod \"nova-api-0\" (UID: \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\") " pod="openstack/nova-api-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.172675 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nwlg\" (UniqueName: \"kubernetes.io/projected/17d3d79f-6c72-4ad5-862e-7a1e020333ae-kube-api-access-5nwlg\") pod \"nova-api-0\" (UID: \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\") " pod="openstack/nova-api-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.172731 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17d3d79f-6c72-4ad5-862e-7a1e020333ae-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\") " pod="openstack/nova-api-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.172806 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17d3d79f-6c72-4ad5-862e-7a1e020333ae-logs\") pod \"nova-api-0\" (UID: \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\") " pod="openstack/nova-api-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.173160 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17d3d79f-6c72-4ad5-862e-7a1e020333ae-logs\") pod \"nova-api-0\" (UID: \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\") " pod="openstack/nova-api-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.175483 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.190768 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17d3d79f-6c72-4ad5-862e-7a1e020333ae-config-data\") pod \"nova-api-0\" (UID: \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\") " pod="openstack/nova-api-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.192692 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17d3d79f-6c72-4ad5-862e-7a1e020333ae-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\") " pod="openstack/nova-api-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.219513 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.219733 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nwlg\" (UniqueName: \"kubernetes.io/projected/17d3d79f-6c72-4ad5-862e-7a1e020333ae-kube-api-access-5nwlg\") pod \"nova-api-0\" (UID: \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\") " pod="openstack/nova-api-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.277989 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dzqm\" (UniqueName: \"kubernetes.io/projected/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-kube-api-access-6dzqm\") pod \"nova-metadata-0\" (UID: \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\") " pod="openstack/nova-metadata-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.278032 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-logs\") pod \"nova-metadata-0\" (UID: \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\") " pod="openstack/nova-metadata-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.278096 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-config-data\") pod \"nova-metadata-0\" (UID: \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\") " pod="openstack/nova-metadata-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.278174 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\") " pod="openstack/nova-metadata-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.296873 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.304902 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-9qq9d"] Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.306517 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.364135 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-9qq9d"] Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.381004 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-9qq9d\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.384204 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-9qq9d\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.384271 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-config\") pod \"dnsmasq-dns-757b4f8459-9qq9d\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.384299 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vltq\" (UniqueName: \"kubernetes.io/projected/37dfca9c-dcd2-4620-b6c6-55225fc0693a-kube-api-access-8vltq\") pod \"dnsmasq-dns-757b4f8459-9qq9d\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.385070 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-dns-svc\") pod \"dnsmasq-dns-757b4f8459-9qq9d\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.385119 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\") " pod="openstack/nova-metadata-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.385225 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-9qq9d\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.385248 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dzqm\" (UniqueName: \"kubernetes.io/projected/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-kube-api-access-6dzqm\") pod \"nova-metadata-0\" (UID: \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\") " pod="openstack/nova-metadata-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.385274 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-logs\") pod \"nova-metadata-0\" (UID: \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\") " pod="openstack/nova-metadata-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.385723 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-logs\") pod \"nova-metadata-0\" (UID: \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\") " pod="openstack/nova-metadata-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.385724 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-config-data\") pod \"nova-metadata-0\" (UID: \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\") " pod="openstack/nova-metadata-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.390555 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.391917 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.394146 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.409807 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-config-data\") pod \"nova-metadata-0\" (UID: \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\") " pod="openstack/nova-metadata-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.411672 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\") " pod="openstack/nova-metadata-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.429548 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.472990 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dzqm\" (UniqueName: \"kubernetes.io/projected/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-kube-api-access-6dzqm\") pod \"nova-metadata-0\" (UID: \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\") " pod="openstack/nova-metadata-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.487774 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-config\") pod \"dnsmasq-dns-757b4f8459-9qq9d\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.487827 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vltq\" (UniqueName: \"kubernetes.io/projected/37dfca9c-dcd2-4620-b6c6-55225fc0693a-kube-api-access-8vltq\") pod \"dnsmasq-dns-757b4f8459-9qq9d\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.487890 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/905037ac-d5f7-46f8-b0ce-aa2649771221-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"905037ac-d5f7-46f8-b0ce-aa2649771221\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.487911 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwsmg\" (UniqueName: \"kubernetes.io/projected/905037ac-d5f7-46f8-b0ce-aa2649771221-kube-api-access-bwsmg\") pod \"nova-cell1-novncproxy-0\" (UID: \"905037ac-d5f7-46f8-b0ce-aa2649771221\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.487935 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-dns-svc\") pod \"dnsmasq-dns-757b4f8459-9qq9d\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.487985 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/905037ac-d5f7-46f8-b0ce-aa2649771221-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"905037ac-d5f7-46f8-b0ce-aa2649771221\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.488013 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-9qq9d\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.488081 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-9qq9d\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.488100 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-9qq9d\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.489564 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-config\") pod \"dnsmasq-dns-757b4f8459-9qq9d\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.490072 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-dns-svc\") pod \"dnsmasq-dns-757b4f8459-9qq9d\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.490260 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-9qq9d\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.490702 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-9qq9d\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.493545 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-9qq9d\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.516706 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vltq\" (UniqueName: \"kubernetes.io/projected/37dfca9c-dcd2-4620-b6c6-55225fc0693a-kube-api-access-8vltq\") pod \"dnsmasq-dns-757b4f8459-9qq9d\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.523649 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.590915 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/905037ac-d5f7-46f8-b0ce-aa2649771221-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"905037ac-d5f7-46f8-b0ce-aa2649771221\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.590972 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwsmg\" (UniqueName: \"kubernetes.io/projected/905037ac-d5f7-46f8-b0ce-aa2649771221-kube-api-access-bwsmg\") pod \"nova-cell1-novncproxy-0\" (UID: \"905037ac-d5f7-46f8-b0ce-aa2649771221\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.591043 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/905037ac-d5f7-46f8-b0ce-aa2649771221-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"905037ac-d5f7-46f8-b0ce-aa2649771221\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.596707 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/905037ac-d5f7-46f8-b0ce-aa2649771221-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"905037ac-d5f7-46f8-b0ce-aa2649771221\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.597178 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/905037ac-d5f7-46f8-b0ce-aa2649771221-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"905037ac-d5f7-46f8-b0ce-aa2649771221\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.626550 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwsmg\" (UniqueName: \"kubernetes.io/projected/905037ac-d5f7-46f8-b0ce-aa2649771221-kube-api-access-bwsmg\") pod \"nova-cell1-novncproxy-0\" (UID: \"905037ac-d5f7-46f8-b0ce-aa2649771221\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.755388 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3169126-e313-4d31-8984-34e321aa0bde","Type":"ContainerStarted","Data":"1d5ed9aad79c96147a4d24724ac5f8e70224f7b6336846171a4811621b800cc1"} Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.765821 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.833464 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:01 crc kubenswrapper[4904]: I0214 11:29:01.899098 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-98lq8"] Feb 14 11:29:01 crc kubenswrapper[4904]: W0214 11:29:01.908401 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod806d3bba_5958_4939_90a5_ae94621e1c5a.slice/crio-208a0c1d55b9c6e145c298b1e079c4c1fd01c13285f990e2a14e63ffcf575cff WatchSource:0}: Error finding container 208a0c1d55b9c6e145c298b1e079c4c1fd01c13285f990e2a14e63ffcf575cff: Status 404 returned error can't find the container with id 208a0c1d55b9c6e145c298b1e079c4c1fd01c13285f990e2a14e63ffcf575cff Feb 14 11:29:02 crc kubenswrapper[4904]: I0214 11:29:02.084821 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 14 11:29:02 crc kubenswrapper[4904]: I0214 11:29:02.300944 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 14 11:29:02 crc kubenswrapper[4904]: I0214 11:29:02.331396 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-9qq9d"] Feb 14 11:29:02 crc kubenswrapper[4904]: I0214 11:29:02.532566 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 14 11:29:02 crc kubenswrapper[4904]: I0214 11:29:02.556190 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:29:02 crc kubenswrapper[4904]: I0214 11:29:02.765098 4904 generic.go:334] "Generic (PLEG): container finished" podID="37dfca9c-dcd2-4620-b6c6-55225fc0693a" containerID="87c70ae353a105453d588f5c729f8c36d0bd7c3717964833db4e1d7e12a5b8fe" exitCode=0 Feb 14 11:29:02 crc kubenswrapper[4904]: I0214 11:29:02.765183 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" event={"ID":"37dfca9c-dcd2-4620-b6c6-55225fc0693a","Type":"ContainerDied","Data":"87c70ae353a105453d588f5c729f8c36d0bd7c3717964833db4e1d7e12a5b8fe"} Feb 14 11:29:02 crc kubenswrapper[4904]: I0214 11:29:02.766350 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" event={"ID":"37dfca9c-dcd2-4620-b6c6-55225fc0693a","Type":"ContainerStarted","Data":"6db0443bff5f272d2c25dcaa444ad3f6f8350254aaca89108fb194ddb152e1c1"} Feb 14 11:29:02 crc kubenswrapper[4904]: I0214 11:29:02.776541 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-98lq8" event={"ID":"806d3bba-5958-4939-90a5-ae94621e1c5a","Type":"ContainerStarted","Data":"be4ad0665d650aa5adb7d3d177de2d9743d103e82015f090cd8312b792aa6e44"} Feb 14 11:29:02 crc kubenswrapper[4904]: I0214 11:29:02.776580 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-98lq8" event={"ID":"806d3bba-5958-4939-90a5-ae94621e1c5a","Type":"ContainerStarted","Data":"208a0c1d55b9c6e145c298b1e079c4c1fd01c13285f990e2a14e63ffcf575cff"} Feb 14 11:29:02 crc kubenswrapper[4904]: I0214 11:29:02.793161 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"17d3d79f-6c72-4ad5-862e-7a1e020333ae","Type":"ContainerStarted","Data":"ad0b20c3b709bd40b259990c265a99d903db27b5bc60f59d4b16dad628f52288"} Feb 14 11:29:02 crc kubenswrapper[4904]: I0214 11:29:02.798127 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c97027f1-aa8d-4457-a4f9-31812b8107d3","Type":"ContainerStarted","Data":"5eec6e1d99857c1f2aeae8ff830dd6f2e27d9f8a3daceb05de765fbe66118fc6"} Feb 14 11:29:02 crc kubenswrapper[4904]: I0214 11:29:02.824034 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"905037ac-d5f7-46f8-b0ce-aa2649771221","Type":"ContainerStarted","Data":"dc40692325f5a3fb4b212d427cb4ea383d11d6cf7ce146d2b45187d8b80bf76b"} Feb 14 11:29:02 crc kubenswrapper[4904]: I0214 11:29:02.833988 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15","Type":"ContainerStarted","Data":"aff97dcefac723af0ef4d754aa3713d15776980d38f2c196ca6f089e7a1ea17c"} Feb 14 11:29:02 crc kubenswrapper[4904]: I0214 11:29:02.851232 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3169126-e313-4d31-8984-34e321aa0bde","Type":"ContainerStarted","Data":"60a52b88d53fc54434a37fb8d8c53c5875d31f5b8923fb2285756021aa4e4c18"} Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.009465 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-98lq8" podStartSLOduration=3.009447654 podStartE2EDuration="3.009447654s" podCreationTimestamp="2026-02-14 11:29:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:29:02.827450329 +0000 UTC m=+1133.640215010" watchObservedRunningTime="2026-02-14 11:29:03.009447654 +0000 UTC m=+1133.822212315" Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.053669 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-nvrps"] Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.057522 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-nvrps" Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.066344 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-nvrps"] Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.069896 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.071250 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.137908 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3011b1dc-9d5e-4c28-bdf6-de2199704f42-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-nvrps\" (UID: \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\") " pod="openstack/nova-cell1-conductor-db-sync-nvrps" Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.138035 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3011b1dc-9d5e-4c28-bdf6-de2199704f42-scripts\") pod \"nova-cell1-conductor-db-sync-nvrps\" (UID: \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\") " pod="openstack/nova-cell1-conductor-db-sync-nvrps" Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.138123 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcjqw\" (UniqueName: \"kubernetes.io/projected/3011b1dc-9d5e-4c28-bdf6-de2199704f42-kube-api-access-mcjqw\") pod \"nova-cell1-conductor-db-sync-nvrps\" (UID: \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\") " pod="openstack/nova-cell1-conductor-db-sync-nvrps" Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.138170 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3011b1dc-9d5e-4c28-bdf6-de2199704f42-config-data\") pod \"nova-cell1-conductor-db-sync-nvrps\" (UID: \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\") " pod="openstack/nova-cell1-conductor-db-sync-nvrps" Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.239812 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3011b1dc-9d5e-4c28-bdf6-de2199704f42-config-data\") pod \"nova-cell1-conductor-db-sync-nvrps\" (UID: \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\") " pod="openstack/nova-cell1-conductor-db-sync-nvrps" Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.240253 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3011b1dc-9d5e-4c28-bdf6-de2199704f42-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-nvrps\" (UID: \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\") " pod="openstack/nova-cell1-conductor-db-sync-nvrps" Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.240393 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3011b1dc-9d5e-4c28-bdf6-de2199704f42-scripts\") pod \"nova-cell1-conductor-db-sync-nvrps\" (UID: \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\") " pod="openstack/nova-cell1-conductor-db-sync-nvrps" Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.240522 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcjqw\" (UniqueName: \"kubernetes.io/projected/3011b1dc-9d5e-4c28-bdf6-de2199704f42-kube-api-access-mcjqw\") pod \"nova-cell1-conductor-db-sync-nvrps\" (UID: \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\") " pod="openstack/nova-cell1-conductor-db-sync-nvrps" Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.248932 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3011b1dc-9d5e-4c28-bdf6-de2199704f42-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-nvrps\" (UID: \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\") " pod="openstack/nova-cell1-conductor-db-sync-nvrps" Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.250004 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3011b1dc-9d5e-4c28-bdf6-de2199704f42-scripts\") pod \"nova-cell1-conductor-db-sync-nvrps\" (UID: \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\") " pod="openstack/nova-cell1-conductor-db-sync-nvrps" Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.266398 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcjqw\" (UniqueName: \"kubernetes.io/projected/3011b1dc-9d5e-4c28-bdf6-de2199704f42-kube-api-access-mcjqw\") pod \"nova-cell1-conductor-db-sync-nvrps\" (UID: \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\") " pod="openstack/nova-cell1-conductor-db-sync-nvrps" Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.270175 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3011b1dc-9d5e-4c28-bdf6-de2199704f42-config-data\") pod \"nova-cell1-conductor-db-sync-nvrps\" (UID: \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\") " pod="openstack/nova-cell1-conductor-db-sync-nvrps" Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.392569 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-nvrps" Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.872282 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" event={"ID":"37dfca9c-dcd2-4620-b6c6-55225fc0693a","Type":"ContainerStarted","Data":"8bbdd4200ba9f703d1be47f0df0a36df1548e458965b28c4ef87d2a9757f4ec7"} Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.872646 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.900097 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" podStartSLOduration=2.9000763750000003 podStartE2EDuration="2.900076375s" podCreationTimestamp="2026-02-14 11:29:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:29:03.892150547 +0000 UTC m=+1134.704915208" watchObservedRunningTime="2026-02-14 11:29:03.900076375 +0000 UTC m=+1134.712841036" Feb 14 11:29:03 crc kubenswrapper[4904]: I0214 11:29:03.983767 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-nvrps"] Feb 14 11:29:04 crc kubenswrapper[4904]: W0214 11:29:04.004271 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3011b1dc_9d5e_4c28_bdf6_de2199704f42.slice/crio-66052b82819dc90c20869a6b0382af2c7278a81b3f43eb64221840ab35e66fc6 WatchSource:0}: Error finding container 66052b82819dc90c20869a6b0382af2c7278a81b3f43eb64221840ab35e66fc6: Status 404 returned error can't find the container with id 66052b82819dc90c20869a6b0382af2c7278a81b3f43eb64221840ab35e66fc6 Feb 14 11:29:04 crc kubenswrapper[4904]: I0214 11:29:04.459384 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 14 11:29:04 crc kubenswrapper[4904]: I0214 11:29:04.465313 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:29:04 crc kubenswrapper[4904]: I0214 11:29:04.898085 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3169126-e313-4d31-8984-34e321aa0bde","Type":"ContainerStarted","Data":"61e971e371cdfdc3d16eb2223dd330107dd8f408c1ad5d7a954aae8ae2346fdb"} Feb 14 11:29:04 crc kubenswrapper[4904]: I0214 11:29:04.898615 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 14 11:29:04 crc kubenswrapper[4904]: I0214 11:29:04.901492 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-nvrps" event={"ID":"3011b1dc-9d5e-4c28-bdf6-de2199704f42","Type":"ContainerStarted","Data":"2ceedd3fe499939062c581b036d4eea72ba156e02350e396a4a38a7e2918a2bd"} Feb 14 11:29:04 crc kubenswrapper[4904]: I0214 11:29:04.901565 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-nvrps" event={"ID":"3011b1dc-9d5e-4c28-bdf6-de2199704f42","Type":"ContainerStarted","Data":"66052b82819dc90c20869a6b0382af2c7278a81b3f43eb64221840ab35e66fc6"} Feb 14 11:29:04 crc kubenswrapper[4904]: I0214 11:29:04.957730 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.201515959 podStartE2EDuration="5.957709968s" podCreationTimestamp="2026-02-14 11:28:59 +0000 UTC" firstStartedPulling="2026-02-14 11:28:59.92365104 +0000 UTC m=+1130.736415701" lastFinishedPulling="2026-02-14 11:29:03.679845049 +0000 UTC m=+1134.492609710" observedRunningTime="2026-02-14 11:29:04.925434491 +0000 UTC m=+1135.738199152" watchObservedRunningTime="2026-02-14 11:29:04.957709968 +0000 UTC m=+1135.770474629" Feb 14 11:29:04 crc kubenswrapper[4904]: I0214 11:29:04.979276 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-nvrps" podStartSLOduration=2.978921051 podStartE2EDuration="2.978921051s" podCreationTimestamp="2026-02-14 11:29:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:29:04.960417573 +0000 UTC m=+1135.773182224" watchObservedRunningTime="2026-02-14 11:29:04.978921051 +0000 UTC m=+1135.791685712" Feb 14 11:29:07 crc kubenswrapper[4904]: I0214 11:29:07.930044 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"905037ac-d5f7-46f8-b0ce-aa2649771221","Type":"ContainerStarted","Data":"5e6f50efb3246fc04bd274dc7fa75e9387dd28c5ea36f158cf11a23757d93c10"} Feb 14 11:29:07 crc kubenswrapper[4904]: I0214 11:29:07.930653 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="905037ac-d5f7-46f8-b0ce-aa2649771221" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://5e6f50efb3246fc04bd274dc7fa75e9387dd28c5ea36f158cf11a23757d93c10" gracePeriod=30 Feb 14 11:29:07 crc kubenswrapper[4904]: I0214 11:29:07.935601 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15","Type":"ContainerStarted","Data":"7a0c207bba00f8b4d5ef34c28c4dca24db9bfb64ea4d7998e5a8bc1f503d0e8d"} Feb 14 11:29:07 crc kubenswrapper[4904]: I0214 11:29:07.935655 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15","Type":"ContainerStarted","Data":"7b09413ebf4351f5c41e8b4116308fc5e094dcb5844973cadd3b4af6776bd444"} Feb 14 11:29:07 crc kubenswrapper[4904]: I0214 11:29:07.935721 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2f259604-d73c-4cfd-b9d9-cdedbbfcdd15" containerName="nova-metadata-log" containerID="cri-o://7b09413ebf4351f5c41e8b4116308fc5e094dcb5844973cadd3b4af6776bd444" gracePeriod=30 Feb 14 11:29:07 crc kubenswrapper[4904]: I0214 11:29:07.935725 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2f259604-d73c-4cfd-b9d9-cdedbbfcdd15" containerName="nova-metadata-metadata" containerID="cri-o://7a0c207bba00f8b4d5ef34c28c4dca24db9bfb64ea4d7998e5a8bc1f503d0e8d" gracePeriod=30 Feb 14 11:29:07 crc kubenswrapper[4904]: I0214 11:29:07.937169 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"17d3d79f-6c72-4ad5-862e-7a1e020333ae","Type":"ContainerStarted","Data":"8f05e84f5e134afd3c686d05171e1ea3c6c237b098109d8dd1f767f57a1224af"} Feb 14 11:29:07 crc kubenswrapper[4904]: I0214 11:29:07.937196 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"17d3d79f-6c72-4ad5-862e-7a1e020333ae","Type":"ContainerStarted","Data":"5345ff80e0c415277499d8c6fda6bae4aafcadad8c75412238ce591e76bd0ea6"} Feb 14 11:29:07 crc kubenswrapper[4904]: I0214 11:29:07.938242 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c97027f1-aa8d-4457-a4f9-31812b8107d3","Type":"ContainerStarted","Data":"993c53a048dccfdb8658894e792ceda74488c299bcef3ba041508c31ec4d5811"} Feb 14 11:29:07 crc kubenswrapper[4904]: I0214 11:29:07.951705 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.867667233 podStartE2EDuration="6.951686767s" podCreationTimestamp="2026-02-14 11:29:01 +0000 UTC" firstStartedPulling="2026-02-14 11:29:02.540221461 +0000 UTC m=+1133.352986122" lastFinishedPulling="2026-02-14 11:29:06.624240995 +0000 UTC m=+1137.437005656" observedRunningTime="2026-02-14 11:29:07.946483383 +0000 UTC m=+1138.759248045" watchObservedRunningTime="2026-02-14 11:29:07.951686767 +0000 UTC m=+1138.764451428" Feb 14 11:29:07 crc kubenswrapper[4904]: I0214 11:29:07.974485 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.664346473 podStartE2EDuration="7.974470773s" podCreationTimestamp="2026-02-14 11:29:00 +0000 UTC" firstStartedPulling="2026-02-14 11:29:02.315257326 +0000 UTC m=+1133.128021987" lastFinishedPulling="2026-02-14 11:29:06.625381626 +0000 UTC m=+1137.438146287" observedRunningTime="2026-02-14 11:29:07.97288656 +0000 UTC m=+1138.785651221" watchObservedRunningTime="2026-02-14 11:29:07.974470773 +0000 UTC m=+1138.787235434" Feb 14 11:29:08 crc kubenswrapper[4904]: I0214 11:29:08.018779 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.874638546 podStartE2EDuration="7.018760761s" podCreationTimestamp="2026-02-14 11:29:01 +0000 UTC" firstStartedPulling="2026-02-14 11:29:02.582576577 +0000 UTC m=+1133.395341238" lastFinishedPulling="2026-02-14 11:29:06.726698792 +0000 UTC m=+1137.539463453" observedRunningTime="2026-02-14 11:29:08.002970148 +0000 UTC m=+1138.815734809" watchObservedRunningTime="2026-02-14 11:29:08.018760761 +0000 UTC m=+1138.831525422" Feb 14 11:29:08 crc kubenswrapper[4904]: I0214 11:29:08.019405 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.497541956 podStartE2EDuration="8.01939911s" podCreationTimestamp="2026-02-14 11:29:00 +0000 UTC" firstStartedPulling="2026-02-14 11:29:02.100301594 +0000 UTC m=+1132.913066255" lastFinishedPulling="2026-02-14 11:29:06.622158748 +0000 UTC m=+1137.434923409" observedRunningTime="2026-02-14 11:29:08.016254872 +0000 UTC m=+1138.829019533" watchObservedRunningTime="2026-02-14 11:29:08.01939911 +0000 UTC m=+1138.832163771" Feb 14 11:29:08 crc kubenswrapper[4904]: I0214 11:29:08.948690 4904 generic.go:334] "Generic (PLEG): container finished" podID="2f259604-d73c-4cfd-b9d9-cdedbbfcdd15" containerID="7a0c207bba00f8b4d5ef34c28c4dca24db9bfb64ea4d7998e5a8bc1f503d0e8d" exitCode=0 Feb 14 11:29:08 crc kubenswrapper[4904]: I0214 11:29:08.949040 4904 generic.go:334] "Generic (PLEG): container finished" podID="2f259604-d73c-4cfd-b9d9-cdedbbfcdd15" containerID="7b09413ebf4351f5c41e8b4116308fc5e094dcb5844973cadd3b4af6776bd444" exitCode=143 Feb 14 11:29:08 crc kubenswrapper[4904]: I0214 11:29:08.948736 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15","Type":"ContainerDied","Data":"7a0c207bba00f8b4d5ef34c28c4dca24db9bfb64ea4d7998e5a8bc1f503d0e8d"} Feb 14 11:29:08 crc kubenswrapper[4904]: I0214 11:29:08.949127 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15","Type":"ContainerDied","Data":"7b09413ebf4351f5c41e8b4116308fc5e094dcb5844973cadd3b4af6776bd444"} Feb 14 11:29:08 crc kubenswrapper[4904]: I0214 11:29:08.949143 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15","Type":"ContainerDied","Data":"aff97dcefac723af0ef4d754aa3713d15776980d38f2c196ca6f089e7a1ea17c"} Feb 14 11:29:08 crc kubenswrapper[4904]: I0214 11:29:08.949155 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aff97dcefac723af0ef4d754aa3713d15776980d38f2c196ca6f089e7a1ea17c" Feb 14 11:29:08 crc kubenswrapper[4904]: I0214 11:29:08.949379 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-687657b496-tz596" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 14 11:29:09 crc kubenswrapper[4904]: I0214 11:29:09.018711 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 14 11:29:09 crc kubenswrapper[4904]: I0214 11:29:09.087451 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dzqm\" (UniqueName: \"kubernetes.io/projected/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-kube-api-access-6dzqm\") pod \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\" (UID: \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\") " Feb 14 11:29:09 crc kubenswrapper[4904]: I0214 11:29:09.087620 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-config-data\") pod \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\" (UID: \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\") " Feb 14 11:29:09 crc kubenswrapper[4904]: I0214 11:29:09.087656 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-logs\") pod \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\" (UID: \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\") " Feb 14 11:29:09 crc kubenswrapper[4904]: I0214 11:29:09.087704 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-combined-ca-bundle\") pod \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\" (UID: \"2f259604-d73c-4cfd-b9d9-cdedbbfcdd15\") " Feb 14 11:29:09 crc kubenswrapper[4904]: I0214 11:29:09.093296 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-logs" (OuterVolumeSpecName: "logs") pod "2f259604-d73c-4cfd-b9d9-cdedbbfcdd15" (UID: "2f259604-d73c-4cfd-b9d9-cdedbbfcdd15"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:29:09 crc kubenswrapper[4904]: I0214 11:29:09.129973 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f259604-d73c-4cfd-b9d9-cdedbbfcdd15" (UID: "2f259604-d73c-4cfd-b9d9-cdedbbfcdd15"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:09 crc kubenswrapper[4904]: I0214 11:29:09.135091 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-kube-api-access-6dzqm" (OuterVolumeSpecName: "kube-api-access-6dzqm") pod "2f259604-d73c-4cfd-b9d9-cdedbbfcdd15" (UID: "2f259604-d73c-4cfd-b9d9-cdedbbfcdd15"). InnerVolumeSpecName "kube-api-access-6dzqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:29:09 crc kubenswrapper[4904]: I0214 11:29:09.167973 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-config-data" (OuterVolumeSpecName: "config-data") pod "2f259604-d73c-4cfd-b9d9-cdedbbfcdd15" (UID: "2f259604-d73c-4cfd-b9d9-cdedbbfcdd15"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:09 crc kubenswrapper[4904]: I0214 11:29:09.189909 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dzqm\" (UniqueName: \"kubernetes.io/projected/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-kube-api-access-6dzqm\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:09 crc kubenswrapper[4904]: I0214 11:29:09.189943 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:09 crc kubenswrapper[4904]: I0214 11:29:09.189951 4904 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-logs\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:09 crc kubenswrapper[4904]: I0214 11:29:09.189959 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:09 crc kubenswrapper[4904]: I0214 11:29:09.958634 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 14 11:29:09 crc kubenswrapper[4904]: I0214 11:29:09.991046 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.021167 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.056177 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:29:10 crc kubenswrapper[4904]: E0214 11:29:10.056970 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f259604-d73c-4cfd-b9d9-cdedbbfcdd15" containerName="nova-metadata-log" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.056988 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f259604-d73c-4cfd-b9d9-cdedbbfcdd15" containerName="nova-metadata-log" Feb 14 11:29:10 crc kubenswrapper[4904]: E0214 11:29:10.057017 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f259604-d73c-4cfd-b9d9-cdedbbfcdd15" containerName="nova-metadata-metadata" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.057023 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f259604-d73c-4cfd-b9d9-cdedbbfcdd15" containerName="nova-metadata-metadata" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.057183 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f259604-d73c-4cfd-b9d9-cdedbbfcdd15" containerName="nova-metadata-metadata" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.057198 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f259604-d73c-4cfd-b9d9-cdedbbfcdd15" containerName="nova-metadata-log" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.062386 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.071702 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.071865 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.094723 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.107453 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g22x\" (UniqueName: \"kubernetes.io/projected/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-kube-api-access-8g22x\") pod \"nova-metadata-0\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " pod="openstack/nova-metadata-0" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.107570 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-logs\") pod \"nova-metadata-0\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " pod="openstack/nova-metadata-0" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.107622 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-config-data\") pod \"nova-metadata-0\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " pod="openstack/nova-metadata-0" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.107650 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " pod="openstack/nova-metadata-0" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.107727 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " pod="openstack/nova-metadata-0" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.209725 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g22x\" (UniqueName: \"kubernetes.io/projected/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-kube-api-access-8g22x\") pod \"nova-metadata-0\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " pod="openstack/nova-metadata-0" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.209804 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-logs\") pod \"nova-metadata-0\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " pod="openstack/nova-metadata-0" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.209881 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-config-data\") pod \"nova-metadata-0\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " pod="openstack/nova-metadata-0" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.209908 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " pod="openstack/nova-metadata-0" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.209978 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " pod="openstack/nova-metadata-0" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.210911 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-logs\") pod \"nova-metadata-0\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " pod="openstack/nova-metadata-0" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.214278 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " pod="openstack/nova-metadata-0" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.223501 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-config-data\") pod \"nova-metadata-0\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " pod="openstack/nova-metadata-0" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.224507 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " pod="openstack/nova-metadata-0" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.227790 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g22x\" (UniqueName: \"kubernetes.io/projected/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-kube-api-access-8g22x\") pod \"nova-metadata-0\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " pod="openstack/nova-metadata-0" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.393741 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.903346 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:29:10 crc kubenswrapper[4904]: I0214 11:29:10.968127 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"81cb9041-d0b1-42f6-9414-a44f4f7af8b6","Type":"ContainerStarted","Data":"555f3d3d1d2ac2968e6d6fdfa067aee949feaddad5d111377bb08d6805bfabac"} Feb 14 11:29:11 crc kubenswrapper[4904]: I0214 11:29:11.222260 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 14 11:29:11 crc kubenswrapper[4904]: I0214 11:29:11.222620 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 14 11:29:11 crc kubenswrapper[4904]: I0214 11:29:11.260926 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 14 11:29:11 crc kubenswrapper[4904]: I0214 11:29:11.297628 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 14 11:29:11 crc kubenswrapper[4904]: I0214 11:29:11.297670 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 14 11:29:11 crc kubenswrapper[4904]: I0214 11:29:11.526058 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:11 crc kubenswrapper[4904]: I0214 11:29:11.609253 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-s54q7"] Feb 14 11:29:11 crc kubenswrapper[4904]: I0214 11:29:11.609702 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" podUID="0d09a98a-7e3d-405a-b2d1-6f6537f4c940" containerName="dnsmasq-dns" containerID="cri-o://a0f74c1dcef80c1649018c3d8c6f171033515a1a167abcbd736a974e775268cc" gracePeriod=10 Feb 14 11:29:11 crc kubenswrapper[4904]: I0214 11:29:11.835575 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:11 crc kubenswrapper[4904]: I0214 11:29:11.852557 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f259604-d73c-4cfd-b9d9-cdedbbfcdd15" path="/var/lib/kubelet/pods/2f259604-d73c-4cfd-b9d9-cdedbbfcdd15/volumes" Feb 14 11:29:11 crc kubenswrapper[4904]: I0214 11:29:11.979462 4904 generic.go:334] "Generic (PLEG): container finished" podID="0d09a98a-7e3d-405a-b2d1-6f6537f4c940" containerID="a0f74c1dcef80c1649018c3d8c6f171033515a1a167abcbd736a974e775268cc" exitCode=0 Feb 14 11:29:11 crc kubenswrapper[4904]: I0214 11:29:11.979563 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" event={"ID":"0d09a98a-7e3d-405a-b2d1-6f6537f4c940","Type":"ContainerDied","Data":"a0f74c1dcef80c1649018c3d8c6f171033515a1a167abcbd736a974e775268cc"} Feb 14 11:29:11 crc kubenswrapper[4904]: I0214 11:29:11.985035 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"81cb9041-d0b1-42f6-9414-a44f4f7af8b6","Type":"ContainerStarted","Data":"cea777a0cc52aaa882e37d8006d8bab6688374680aac25e3a57265aab8be7168"} Feb 14 11:29:11 crc kubenswrapper[4904]: I0214 11:29:11.985195 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"81cb9041-d0b1-42f6-9414-a44f4f7af8b6","Type":"ContainerStarted","Data":"81f730014ea4ef4b93aab8b32e060cf1295a752f1135bb8a64dd1d9a043d93c1"} Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.003573 4904 generic.go:334] "Generic (PLEG): container finished" podID="806d3bba-5958-4939-90a5-ae94621e1c5a" containerID="be4ad0665d650aa5adb7d3d177de2d9743d103e82015f090cd8312b792aa6e44" exitCode=0 Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.004391 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-98lq8" event={"ID":"806d3bba-5958-4939-90a5-ae94621e1c5a","Type":"ContainerDied","Data":"be4ad0665d650aa5adb7d3d177de2d9743d103e82015f090cd8312b792aa6e44"} Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.031989 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.031971527 podStartE2EDuration="3.031971527s" podCreationTimestamp="2026-02-14 11:29:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:29:12.027187516 +0000 UTC m=+1142.839952177" watchObservedRunningTime="2026-02-14 11:29:12.031971527 +0000 UTC m=+1142.844736188" Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.071363 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.195781 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.266322 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-ovsdbserver-sb\") pod \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.266709 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-dns-svc\") pod \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.266848 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-ovsdbserver-nb\") pod \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.266958 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-dns-swift-storage-0\") pod \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.267104 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-config\") pod \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.267242 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghwtp\" (UniqueName: \"kubernetes.io/projected/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-kube-api-access-ghwtp\") pod \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.281072 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-kube-api-access-ghwtp" (OuterVolumeSpecName: "kube-api-access-ghwtp") pod "0d09a98a-7e3d-405a-b2d1-6f6537f4c940" (UID: "0d09a98a-7e3d-405a-b2d1-6f6537f4c940"). InnerVolumeSpecName "kube-api-access-ghwtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.353270 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-config" (OuterVolumeSpecName: "config") pod "0d09a98a-7e3d-405a-b2d1-6f6537f4c940" (UID: "0d09a98a-7e3d-405a-b2d1-6f6537f4c940"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.367816 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0d09a98a-7e3d-405a-b2d1-6f6537f4c940" (UID: "0d09a98a-7e3d-405a-b2d1-6f6537f4c940"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.373807 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-ovsdbserver-nb\") pod \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\" (UID: \"0d09a98a-7e3d-405a-b2d1-6f6537f4c940\") " Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.374330 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.374747 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghwtp\" (UniqueName: \"kubernetes.io/projected/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-kube-api-access-ghwtp\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:12 crc kubenswrapper[4904]: W0214 11:29:12.374722 4904 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/0d09a98a-7e3d-405a-b2d1-6f6537f4c940/volumes/kubernetes.io~configmap/ovsdbserver-nb Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.375432 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0d09a98a-7e3d-405a-b2d1-6f6537f4c940" (UID: "0d09a98a-7e3d-405a-b2d1-6f6537f4c940"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.380000 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="17d3d79f-6c72-4ad5-862e-7a1e020333ae" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.380016 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="17d3d79f-6c72-4ad5-862e-7a1e020333ae" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.381309 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0d09a98a-7e3d-405a-b2d1-6f6537f4c940" (UID: "0d09a98a-7e3d-405a-b2d1-6f6537f4c940"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.390674 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0d09a98a-7e3d-405a-b2d1-6f6537f4c940" (UID: "0d09a98a-7e3d-405a-b2d1-6f6537f4c940"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.413317 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0d09a98a-7e3d-405a-b2d1-6f6537f4c940" (UID: "0d09a98a-7e3d-405a-b2d1-6f6537f4c940"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.476418 4904 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.476452 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.476463 4904 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:12 crc kubenswrapper[4904]: I0214 11:29:12.476471 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d09a98a-7e3d-405a-b2d1-6f6537f4c940-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.017200 4904 generic.go:334] "Generic (PLEG): container finished" podID="3011b1dc-9d5e-4c28-bdf6-de2199704f42" containerID="2ceedd3fe499939062c581b036d4eea72ba156e02350e396a4a38a7e2918a2bd" exitCode=0 Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.017535 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-nvrps" event={"ID":"3011b1dc-9d5e-4c28-bdf6-de2199704f42","Type":"ContainerDied","Data":"2ceedd3fe499939062c581b036d4eea72ba156e02350e396a4a38a7e2918a2bd"} Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.021100 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.030049 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-s54q7" event={"ID":"0d09a98a-7e3d-405a-b2d1-6f6537f4c940","Type":"ContainerDied","Data":"84522020ca1948cf4c8dbca5113534cc75e06bf03ae4a2e083e3569883c167e7"} Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.030085 4904 scope.go:117] "RemoveContainer" containerID="a0f74c1dcef80c1649018c3d8c6f171033515a1a167abcbd736a974e775268cc" Feb 14 11:29:13 crc kubenswrapper[4904]: E0214 11:29:13.041930 4904 kuberuntime_gc.go:389] "Failed to remove container log dead symlink" err="remove /var/log/containers/dnsmasq-dns-5c9776ccc5-s54q7_openstack_dnsmasq-dns-a0f74c1dcef80c1649018c3d8c6f171033515a1a167abcbd736a974e775268cc.log: no such file or directory" path="/var/log/containers/dnsmasq-dns-5c9776ccc5-s54q7_openstack_dnsmasq-dns-a0f74c1dcef80c1649018c3d8c6f171033515a1a167abcbd736a974e775268cc.log" Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.061584 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-s54q7"] Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.069726 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-s54q7"] Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.095888 4904 scope.go:117] "RemoveContainer" containerID="73797f3df06ae349b13be6fb4aa32a1c40834ec93034f0999fb191bcfaaa20bd" Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.494501 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-98lq8" Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.596369 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/806d3bba-5958-4939-90a5-ae94621e1c5a-scripts\") pod \"806d3bba-5958-4939-90a5-ae94621e1c5a\" (UID: \"806d3bba-5958-4939-90a5-ae94621e1c5a\") " Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.596429 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/806d3bba-5958-4939-90a5-ae94621e1c5a-combined-ca-bundle\") pod \"806d3bba-5958-4939-90a5-ae94621e1c5a\" (UID: \"806d3bba-5958-4939-90a5-ae94621e1c5a\") " Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.596450 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/806d3bba-5958-4939-90a5-ae94621e1c5a-config-data\") pod \"806d3bba-5958-4939-90a5-ae94621e1c5a\" (UID: \"806d3bba-5958-4939-90a5-ae94621e1c5a\") " Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.596543 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjmvl\" (UniqueName: \"kubernetes.io/projected/806d3bba-5958-4939-90a5-ae94621e1c5a-kube-api-access-kjmvl\") pod \"806d3bba-5958-4939-90a5-ae94621e1c5a\" (UID: \"806d3bba-5958-4939-90a5-ae94621e1c5a\") " Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.625752 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/806d3bba-5958-4939-90a5-ae94621e1c5a-kube-api-access-kjmvl" (OuterVolumeSpecName: "kube-api-access-kjmvl") pod "806d3bba-5958-4939-90a5-ae94621e1c5a" (UID: "806d3bba-5958-4939-90a5-ae94621e1c5a"). InnerVolumeSpecName "kube-api-access-kjmvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.629013 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/806d3bba-5958-4939-90a5-ae94621e1c5a-scripts" (OuterVolumeSpecName: "scripts") pod "806d3bba-5958-4939-90a5-ae94621e1c5a" (UID: "806d3bba-5958-4939-90a5-ae94621e1c5a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.639353 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/806d3bba-5958-4939-90a5-ae94621e1c5a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "806d3bba-5958-4939-90a5-ae94621e1c5a" (UID: "806d3bba-5958-4939-90a5-ae94621e1c5a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.640200 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/806d3bba-5958-4939-90a5-ae94621e1c5a-config-data" (OuterVolumeSpecName: "config-data") pod "806d3bba-5958-4939-90a5-ae94621e1c5a" (UID: "806d3bba-5958-4939-90a5-ae94621e1c5a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.698503 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/806d3bba-5958-4939-90a5-ae94621e1c5a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.698542 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/806d3bba-5958-4939-90a5-ae94621e1c5a-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.698554 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjmvl\" (UniqueName: \"kubernetes.io/projected/806d3bba-5958-4939-90a5-ae94621e1c5a-kube-api-access-kjmvl\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.698570 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/806d3bba-5958-4939-90a5-ae94621e1c5a-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:13 crc kubenswrapper[4904]: I0214 11:29:13.850038 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d09a98a-7e3d-405a-b2d1-6f6537f4c940" path="/var/lib/kubelet/pods/0d09a98a-7e3d-405a-b2d1-6f6537f4c940/volumes" Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.031750 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-98lq8" event={"ID":"806d3bba-5958-4939-90a5-ae94621e1c5a","Type":"ContainerDied","Data":"208a0c1d55b9c6e145c298b1e079c4c1fd01c13285f990e2a14e63ffcf575cff"} Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.031794 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="208a0c1d55b9c6e145c298b1e079c4c1fd01c13285f990e2a14e63ffcf575cff" Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.031763 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-98lq8" Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.246008 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.246275 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="17d3d79f-6c72-4ad5-862e-7a1e020333ae" containerName="nova-api-log" containerID="cri-o://5345ff80e0c415277499d8c6fda6bae4aafcadad8c75412238ce591e76bd0ea6" gracePeriod=30 Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.247899 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="17d3d79f-6c72-4ad5-862e-7a1e020333ae" containerName="nova-api-api" containerID="cri-o://8f05e84f5e134afd3c686d05171e1ea3c6c237b098109d8dd1f767f57a1224af" gracePeriod=30 Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.307157 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.307395 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c97027f1-aa8d-4457-a4f9-31812b8107d3" containerName="nova-scheduler-scheduler" containerID="cri-o://993c53a048dccfdb8658894e792ceda74488c299bcef3ba041508c31ec4d5811" gracePeriod=30 Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.407598 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.408126 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="81cb9041-d0b1-42f6-9414-a44f4f7af8b6" containerName="nova-metadata-log" containerID="cri-o://81f730014ea4ef4b93aab8b32e060cf1295a752f1135bb8a64dd1d9a043d93c1" gracePeriod=30 Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.408258 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="81cb9041-d0b1-42f6-9414-a44f4f7af8b6" containerName="nova-metadata-metadata" containerID="cri-o://cea777a0cc52aaa882e37d8006d8bab6688374680aac25e3a57265aab8be7168" gracePeriod=30 Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.465117 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-nvrps" Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.517588 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3011b1dc-9d5e-4c28-bdf6-de2199704f42-scripts\") pod \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\" (UID: \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\") " Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.517652 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3011b1dc-9d5e-4c28-bdf6-de2199704f42-config-data\") pod \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\" (UID: \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\") " Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.517754 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3011b1dc-9d5e-4c28-bdf6-de2199704f42-combined-ca-bundle\") pod \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\" (UID: \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\") " Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.517818 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcjqw\" (UniqueName: \"kubernetes.io/projected/3011b1dc-9d5e-4c28-bdf6-de2199704f42-kube-api-access-mcjqw\") pod \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\" (UID: \"3011b1dc-9d5e-4c28-bdf6-de2199704f42\") " Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.536949 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3011b1dc-9d5e-4c28-bdf6-de2199704f42-scripts" (OuterVolumeSpecName: "scripts") pod "3011b1dc-9d5e-4c28-bdf6-de2199704f42" (UID: "3011b1dc-9d5e-4c28-bdf6-de2199704f42"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.537051 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3011b1dc-9d5e-4c28-bdf6-de2199704f42-kube-api-access-mcjqw" (OuterVolumeSpecName: "kube-api-access-mcjqw") pod "3011b1dc-9d5e-4c28-bdf6-de2199704f42" (UID: "3011b1dc-9d5e-4c28-bdf6-de2199704f42"). InnerVolumeSpecName "kube-api-access-mcjqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.561651 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3011b1dc-9d5e-4c28-bdf6-de2199704f42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3011b1dc-9d5e-4c28-bdf6-de2199704f42" (UID: "3011b1dc-9d5e-4c28-bdf6-de2199704f42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.595990 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3011b1dc-9d5e-4c28-bdf6-de2199704f42-config-data" (OuterVolumeSpecName: "config-data") pod "3011b1dc-9d5e-4c28-bdf6-de2199704f42" (UID: "3011b1dc-9d5e-4c28-bdf6-de2199704f42"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.623154 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3011b1dc-9d5e-4c28-bdf6-de2199704f42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.623188 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcjqw\" (UniqueName: \"kubernetes.io/projected/3011b1dc-9d5e-4c28-bdf6-de2199704f42-kube-api-access-mcjqw\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.623201 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3011b1dc-9d5e-4c28-bdf6-de2199704f42-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.623209 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3011b1dc-9d5e-4c28-bdf6-de2199704f42-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:14 crc kubenswrapper[4904]: I0214 11:29:14.957486 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.032535 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-config-data\") pod \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.032831 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-nova-metadata-tls-certs\") pod \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.032875 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-combined-ca-bundle\") pod \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.032988 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8g22x\" (UniqueName: \"kubernetes.io/projected/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-kube-api-access-8g22x\") pod \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.033079 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-logs\") pod \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\" (UID: \"81cb9041-d0b1-42f6-9414-a44f4f7af8b6\") " Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.033669 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-logs" (OuterVolumeSpecName: "logs") pod "81cb9041-d0b1-42f6-9414-a44f4f7af8b6" (UID: "81cb9041-d0b1-42f6-9414-a44f4f7af8b6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.043316 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-kube-api-access-8g22x" (OuterVolumeSpecName: "kube-api-access-8g22x") pod "81cb9041-d0b1-42f6-9414-a44f4f7af8b6" (UID: "81cb9041-d0b1-42f6-9414-a44f4f7af8b6"). InnerVolumeSpecName "kube-api-access-8g22x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.086479 4904 generic.go:334] "Generic (PLEG): container finished" podID="81cb9041-d0b1-42f6-9414-a44f4f7af8b6" containerID="cea777a0cc52aaa882e37d8006d8bab6688374680aac25e3a57265aab8be7168" exitCode=0 Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.086508 4904 generic.go:334] "Generic (PLEG): container finished" podID="81cb9041-d0b1-42f6-9414-a44f4f7af8b6" containerID="81f730014ea4ef4b93aab8b32e060cf1295a752f1135bb8a64dd1d9a043d93c1" exitCode=143 Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.086572 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.086601 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"81cb9041-d0b1-42f6-9414-a44f4f7af8b6","Type":"ContainerDied","Data":"cea777a0cc52aaa882e37d8006d8bab6688374680aac25e3a57265aab8be7168"} Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.086642 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"81cb9041-d0b1-42f6-9414-a44f4f7af8b6","Type":"ContainerDied","Data":"81f730014ea4ef4b93aab8b32e060cf1295a752f1135bb8a64dd1d9a043d93c1"} Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.086654 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"81cb9041-d0b1-42f6-9414-a44f4f7af8b6","Type":"ContainerDied","Data":"555f3d3d1d2ac2968e6d6fdfa067aee949feaddad5d111377bb08d6805bfabac"} Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.086668 4904 scope.go:117] "RemoveContainer" containerID="cea777a0cc52aaa882e37d8006d8bab6688374680aac25e3a57265aab8be7168" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.102077 4904 generic.go:334] "Generic (PLEG): container finished" podID="17d3d79f-6c72-4ad5-862e-7a1e020333ae" containerID="5345ff80e0c415277499d8c6fda6bae4aafcadad8c75412238ce591e76bd0ea6" exitCode=143 Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.102134 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"17d3d79f-6c72-4ad5-862e-7a1e020333ae","Type":"ContainerDied","Data":"5345ff80e0c415277499d8c6fda6bae4aafcadad8c75412238ce591e76bd0ea6"} Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.103597 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-nvrps" event={"ID":"3011b1dc-9d5e-4c28-bdf6-de2199704f42","Type":"ContainerDied","Data":"66052b82819dc90c20869a6b0382af2c7278a81b3f43eb64221840ab35e66fc6"} Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.103615 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66052b82819dc90c20869a6b0382af2c7278a81b3f43eb64221840ab35e66fc6" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.103660 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-nvrps" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.123815 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81cb9041-d0b1-42f6-9414-a44f4f7af8b6" (UID: "81cb9041-d0b1-42f6-9414-a44f4f7af8b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.124161 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-config-data" (OuterVolumeSpecName: "config-data") pod "81cb9041-d0b1-42f6-9414-a44f4f7af8b6" (UID: "81cb9041-d0b1-42f6-9414-a44f4f7af8b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.135267 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.135291 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.135301 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8g22x\" (UniqueName: \"kubernetes.io/projected/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-kube-api-access-8g22x\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.135333 4904 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-logs\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.148354 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 14 11:29:15 crc kubenswrapper[4904]: E0214 11:29:15.148802 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81cb9041-d0b1-42f6-9414-a44f4f7af8b6" containerName="nova-metadata-log" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.148819 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="81cb9041-d0b1-42f6-9414-a44f4f7af8b6" containerName="nova-metadata-log" Feb 14 11:29:15 crc kubenswrapper[4904]: E0214 11:29:15.148858 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81cb9041-d0b1-42f6-9414-a44f4f7af8b6" containerName="nova-metadata-metadata" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.148865 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="81cb9041-d0b1-42f6-9414-a44f4f7af8b6" containerName="nova-metadata-metadata" Feb 14 11:29:15 crc kubenswrapper[4904]: E0214 11:29:15.148877 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d09a98a-7e3d-405a-b2d1-6f6537f4c940" containerName="init" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.148883 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d09a98a-7e3d-405a-b2d1-6f6537f4c940" containerName="init" Feb 14 11:29:15 crc kubenswrapper[4904]: E0214 11:29:15.148892 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3011b1dc-9d5e-4c28-bdf6-de2199704f42" containerName="nova-cell1-conductor-db-sync" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.148897 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="3011b1dc-9d5e-4c28-bdf6-de2199704f42" containerName="nova-cell1-conductor-db-sync" Feb 14 11:29:15 crc kubenswrapper[4904]: E0214 11:29:15.148910 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d09a98a-7e3d-405a-b2d1-6f6537f4c940" containerName="dnsmasq-dns" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.148915 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d09a98a-7e3d-405a-b2d1-6f6537f4c940" containerName="dnsmasq-dns" Feb 14 11:29:15 crc kubenswrapper[4904]: E0214 11:29:15.152909 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="806d3bba-5958-4939-90a5-ae94621e1c5a" containerName="nova-manage" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.152933 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="806d3bba-5958-4939-90a5-ae94621e1c5a" containerName="nova-manage" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.155260 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "81cb9041-d0b1-42f6-9414-a44f4f7af8b6" (UID: "81cb9041-d0b1-42f6-9414-a44f4f7af8b6"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.157353 4904 scope.go:117] "RemoveContainer" containerID="81f730014ea4ef4b93aab8b32e060cf1295a752f1135bb8a64dd1d9a043d93c1" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.157646 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="3011b1dc-9d5e-4c28-bdf6-de2199704f42" containerName="nova-cell1-conductor-db-sync" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.157700 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d09a98a-7e3d-405a-b2d1-6f6537f4c940" containerName="dnsmasq-dns" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.157713 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="806d3bba-5958-4939-90a5-ae94621e1c5a" containerName="nova-manage" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.157729 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="81cb9041-d0b1-42f6-9414-a44f4f7af8b6" containerName="nova-metadata-metadata" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.157749 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="81cb9041-d0b1-42f6-9414-a44f4f7af8b6" containerName="nova-metadata-log" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.160319 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.160480 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.162599 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.192647 4904 scope.go:117] "RemoveContainer" containerID="cea777a0cc52aaa882e37d8006d8bab6688374680aac25e3a57265aab8be7168" Feb 14 11:29:15 crc kubenswrapper[4904]: E0214 11:29:15.197276 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cea777a0cc52aaa882e37d8006d8bab6688374680aac25e3a57265aab8be7168\": container with ID starting with cea777a0cc52aaa882e37d8006d8bab6688374680aac25e3a57265aab8be7168 not found: ID does not exist" containerID="cea777a0cc52aaa882e37d8006d8bab6688374680aac25e3a57265aab8be7168" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.197310 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cea777a0cc52aaa882e37d8006d8bab6688374680aac25e3a57265aab8be7168"} err="failed to get container status \"cea777a0cc52aaa882e37d8006d8bab6688374680aac25e3a57265aab8be7168\": rpc error: code = NotFound desc = could not find container \"cea777a0cc52aaa882e37d8006d8bab6688374680aac25e3a57265aab8be7168\": container with ID starting with cea777a0cc52aaa882e37d8006d8bab6688374680aac25e3a57265aab8be7168 not found: ID does not exist" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.197332 4904 scope.go:117] "RemoveContainer" containerID="81f730014ea4ef4b93aab8b32e060cf1295a752f1135bb8a64dd1d9a043d93c1" Feb 14 11:29:15 crc kubenswrapper[4904]: E0214 11:29:15.200255 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81f730014ea4ef4b93aab8b32e060cf1295a752f1135bb8a64dd1d9a043d93c1\": container with ID starting with 81f730014ea4ef4b93aab8b32e060cf1295a752f1135bb8a64dd1d9a043d93c1 not found: ID does not exist" containerID="81f730014ea4ef4b93aab8b32e060cf1295a752f1135bb8a64dd1d9a043d93c1" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.200289 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81f730014ea4ef4b93aab8b32e060cf1295a752f1135bb8a64dd1d9a043d93c1"} err="failed to get container status \"81f730014ea4ef4b93aab8b32e060cf1295a752f1135bb8a64dd1d9a043d93c1\": rpc error: code = NotFound desc = could not find container \"81f730014ea4ef4b93aab8b32e060cf1295a752f1135bb8a64dd1d9a043d93c1\": container with ID starting with 81f730014ea4ef4b93aab8b32e060cf1295a752f1135bb8a64dd1d9a043d93c1 not found: ID does not exist" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.200307 4904 scope.go:117] "RemoveContainer" containerID="cea777a0cc52aaa882e37d8006d8bab6688374680aac25e3a57265aab8be7168" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.200508 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cea777a0cc52aaa882e37d8006d8bab6688374680aac25e3a57265aab8be7168"} err="failed to get container status \"cea777a0cc52aaa882e37d8006d8bab6688374680aac25e3a57265aab8be7168\": rpc error: code = NotFound desc = could not find container \"cea777a0cc52aaa882e37d8006d8bab6688374680aac25e3a57265aab8be7168\": container with ID starting with cea777a0cc52aaa882e37d8006d8bab6688374680aac25e3a57265aab8be7168 not found: ID does not exist" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.200527 4904 scope.go:117] "RemoveContainer" containerID="81f730014ea4ef4b93aab8b32e060cf1295a752f1135bb8a64dd1d9a043d93c1" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.200675 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81f730014ea4ef4b93aab8b32e060cf1295a752f1135bb8a64dd1d9a043d93c1"} err="failed to get container status \"81f730014ea4ef4b93aab8b32e060cf1295a752f1135bb8a64dd1d9a043d93c1\": rpc error: code = NotFound desc = could not find container \"81f730014ea4ef4b93aab8b32e060cf1295a752f1135bb8a64dd1d9a043d93c1\": container with ID starting with 81f730014ea4ef4b93aab8b32e060cf1295a752f1135bb8a64dd1d9a043d93c1 not found: ID does not exist" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.237247 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb230e23-e829-4d14-a0ae-9ed889c6840c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fb230e23-e829-4d14-a0ae-9ed889c6840c\") " pod="openstack/nova-cell1-conductor-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.237328 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb230e23-e829-4d14-a0ae-9ed889c6840c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fb230e23-e829-4d14-a0ae-9ed889c6840c\") " pod="openstack/nova-cell1-conductor-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.237419 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krx22\" (UniqueName: \"kubernetes.io/projected/fb230e23-e829-4d14-a0ae-9ed889c6840c-kube-api-access-krx22\") pod \"nova-cell1-conductor-0\" (UID: \"fb230e23-e829-4d14-a0ae-9ed889c6840c\") " pod="openstack/nova-cell1-conductor-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.237491 4904 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/81cb9041-d0b1-42f6-9414-a44f4f7af8b6-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.339634 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb230e23-e829-4d14-a0ae-9ed889c6840c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fb230e23-e829-4d14-a0ae-9ed889c6840c\") " pod="openstack/nova-cell1-conductor-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.339740 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krx22\" (UniqueName: \"kubernetes.io/projected/fb230e23-e829-4d14-a0ae-9ed889c6840c-kube-api-access-krx22\") pod \"nova-cell1-conductor-0\" (UID: \"fb230e23-e829-4d14-a0ae-9ed889c6840c\") " pod="openstack/nova-cell1-conductor-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.339812 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb230e23-e829-4d14-a0ae-9ed889c6840c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fb230e23-e829-4d14-a0ae-9ed889c6840c\") " pod="openstack/nova-cell1-conductor-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.344271 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb230e23-e829-4d14-a0ae-9ed889c6840c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fb230e23-e829-4d14-a0ae-9ed889c6840c\") " pod="openstack/nova-cell1-conductor-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.344380 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb230e23-e829-4d14-a0ae-9ed889c6840c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fb230e23-e829-4d14-a0ae-9ed889c6840c\") " pod="openstack/nova-cell1-conductor-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.354506 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krx22\" (UniqueName: \"kubernetes.io/projected/fb230e23-e829-4d14-a0ae-9ed889c6840c-kube-api-access-krx22\") pod \"nova-cell1-conductor-0\" (UID: \"fb230e23-e829-4d14-a0ae-9ed889c6840c\") " pod="openstack/nova-cell1-conductor-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.424724 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.442274 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.460060 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.462068 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.465026 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.465247 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.477774 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.512733 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.543731 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b248689-1264-4fac-a77e-60f6494317f3-config-data\") pod \"nova-metadata-0\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " pod="openstack/nova-metadata-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.543783 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b248689-1264-4fac-a77e-60f6494317f3-logs\") pod \"nova-metadata-0\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " pod="openstack/nova-metadata-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.543894 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b248689-1264-4fac-a77e-60f6494317f3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " pod="openstack/nova-metadata-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.543941 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t7gf\" (UniqueName: \"kubernetes.io/projected/5b248689-1264-4fac-a77e-60f6494317f3-kube-api-access-5t7gf\") pod \"nova-metadata-0\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " pod="openstack/nova-metadata-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.543990 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b248689-1264-4fac-a77e-60f6494317f3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " pod="openstack/nova-metadata-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.646910 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b248689-1264-4fac-a77e-60f6494317f3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " pod="openstack/nova-metadata-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.647285 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t7gf\" (UniqueName: \"kubernetes.io/projected/5b248689-1264-4fac-a77e-60f6494317f3-kube-api-access-5t7gf\") pod \"nova-metadata-0\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " pod="openstack/nova-metadata-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.647330 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b248689-1264-4fac-a77e-60f6494317f3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " pod="openstack/nova-metadata-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.647414 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b248689-1264-4fac-a77e-60f6494317f3-config-data\") pod \"nova-metadata-0\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " pod="openstack/nova-metadata-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.647450 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b248689-1264-4fac-a77e-60f6494317f3-logs\") pod \"nova-metadata-0\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " pod="openstack/nova-metadata-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.647962 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b248689-1264-4fac-a77e-60f6494317f3-logs\") pod \"nova-metadata-0\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " pod="openstack/nova-metadata-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.652597 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b248689-1264-4fac-a77e-60f6494317f3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " pod="openstack/nova-metadata-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.670946 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b248689-1264-4fac-a77e-60f6494317f3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " pod="openstack/nova-metadata-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.671648 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b248689-1264-4fac-a77e-60f6494317f3-config-data\") pod \"nova-metadata-0\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " pod="openstack/nova-metadata-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.672734 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t7gf\" (UniqueName: \"kubernetes.io/projected/5b248689-1264-4fac-a77e-60f6494317f3-kube-api-access-5t7gf\") pod \"nova-metadata-0\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " pod="openstack/nova-metadata-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.825278 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 14 11:29:15 crc kubenswrapper[4904]: I0214 11:29:15.852006 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81cb9041-d0b1-42f6-9414-a44f4f7af8b6" path="/var/lib/kubelet/pods/81cb9041-d0b1-42f6-9414-a44f4f7af8b6/volumes" Feb 14 11:29:16 crc kubenswrapper[4904]: I0214 11:29:16.017894 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 14 11:29:16 crc kubenswrapper[4904]: I0214 11:29:16.123286 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"fb230e23-e829-4d14-a0ae-9ed889c6840c","Type":"ContainerStarted","Data":"e556aedfa3000485156b06dfd194b8e3ac8211e4e93f22bd14fccd12ceec7aa8"} Feb 14 11:29:16 crc kubenswrapper[4904]: E0214 11:29:16.222831 4904 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="993c53a048dccfdb8658894e792ceda74488c299bcef3ba041508c31ec4d5811" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 14 11:29:16 crc kubenswrapper[4904]: E0214 11:29:16.226291 4904 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="993c53a048dccfdb8658894e792ceda74488c299bcef3ba041508c31ec4d5811" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 14 11:29:16 crc kubenswrapper[4904]: E0214 11:29:16.227475 4904 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="993c53a048dccfdb8658894e792ceda74488c299bcef3ba041508c31ec4d5811" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 14 11:29:16 crc kubenswrapper[4904]: E0214 11:29:16.227513 4904 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c97027f1-aa8d-4457-a4f9-31812b8107d3" containerName="nova-scheduler-scheduler" Feb 14 11:29:16 crc kubenswrapper[4904]: W0214 11:29:16.302332 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b248689_1264_4fac_a77e_60f6494317f3.slice/crio-1c7bfe4372346aceb055b05185212db7e067b8ed80075647c60fb22c06f124bc WatchSource:0}: Error finding container 1c7bfe4372346aceb055b05185212db7e067b8ed80075647c60fb22c06f124bc: Status 404 returned error can't find the container with id 1c7bfe4372346aceb055b05185212db7e067b8ed80075647c60fb22c06f124bc Feb 14 11:29:16 crc kubenswrapper[4904]: I0214 11:29:16.302825 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:29:16 crc kubenswrapper[4904]: I0214 11:29:16.383468 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:29:16 crc kubenswrapper[4904]: I0214 11:29:16.383534 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:29:17 crc kubenswrapper[4904]: I0214 11:29:17.186771 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"fb230e23-e829-4d14-a0ae-9ed889c6840c","Type":"ContainerStarted","Data":"b50d5206bce7da320498c263374aa88ac4f12505648dd5caf172e600e686b2bf"} Feb 14 11:29:17 crc kubenswrapper[4904]: I0214 11:29:17.188273 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 14 11:29:17 crc kubenswrapper[4904]: I0214 11:29:17.206124 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b248689-1264-4fac-a77e-60f6494317f3","Type":"ContainerStarted","Data":"929b4ee3de69c4956fb06ed9107d192c5bc4eada12f4ccf24ecff2018abc128f"} Feb 14 11:29:17 crc kubenswrapper[4904]: I0214 11:29:17.206178 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b248689-1264-4fac-a77e-60f6494317f3","Type":"ContainerStarted","Data":"68b584bdd5beda2ef787496cdb9858bdb40b347ec9054a909d3342040e3d2234"} Feb 14 11:29:17 crc kubenswrapper[4904]: I0214 11:29:17.206188 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b248689-1264-4fac-a77e-60f6494317f3","Type":"ContainerStarted","Data":"1c7bfe4372346aceb055b05185212db7e067b8ed80075647c60fb22c06f124bc"} Feb 14 11:29:17 crc kubenswrapper[4904]: I0214 11:29:17.211300 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.211281429 podStartE2EDuration="2.211281429s" podCreationTimestamp="2026-02-14 11:29:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:29:17.207255049 +0000 UTC m=+1148.020019710" watchObservedRunningTime="2026-02-14 11:29:17.211281429 +0000 UTC m=+1148.024046080" Feb 14 11:29:17 crc kubenswrapper[4904]: I0214 11:29:17.236403 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.236383199 podStartE2EDuration="2.236383199s" podCreationTimestamp="2026-02-14 11:29:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:29:17.223771152 +0000 UTC m=+1148.036535813" watchObservedRunningTime="2026-02-14 11:29:17.236383199 +0000 UTC m=+1148.049147860" Feb 14 11:29:18 crc kubenswrapper[4904]: I0214 11:29:18.949137 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-687657b496-tz596" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Feb 14 11:29:18 crc kubenswrapper[4904]: I0214 11:29:18.949670 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-687657b496-tz596" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.159128 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.219168 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17d3d79f-6c72-4ad5-862e-7a1e020333ae-combined-ca-bundle\") pod \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\" (UID: \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\") " Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.219244 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17d3d79f-6c72-4ad5-862e-7a1e020333ae-logs\") pod \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\" (UID: \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\") " Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.219329 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17d3d79f-6c72-4ad5-862e-7a1e020333ae-config-data\") pod \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\" (UID: \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\") " Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.219399 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nwlg\" (UniqueName: \"kubernetes.io/projected/17d3d79f-6c72-4ad5-862e-7a1e020333ae-kube-api-access-5nwlg\") pod \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\" (UID: \"17d3d79f-6c72-4ad5-862e-7a1e020333ae\") " Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.219753 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17d3d79f-6c72-4ad5-862e-7a1e020333ae-logs" (OuterVolumeSpecName: "logs") pod "17d3d79f-6c72-4ad5-862e-7a1e020333ae" (UID: "17d3d79f-6c72-4ad5-862e-7a1e020333ae"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.224883 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17d3d79f-6c72-4ad5-862e-7a1e020333ae-kube-api-access-5nwlg" (OuterVolumeSpecName: "kube-api-access-5nwlg") pod "17d3d79f-6c72-4ad5-862e-7a1e020333ae" (UID: "17d3d79f-6c72-4ad5-862e-7a1e020333ae"). InnerVolumeSpecName "kube-api-access-5nwlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.241213 4904 generic.go:334] "Generic (PLEG): container finished" podID="c97027f1-aa8d-4457-a4f9-31812b8107d3" containerID="993c53a048dccfdb8658894e792ceda74488c299bcef3ba041508c31ec4d5811" exitCode=0 Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.241275 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c97027f1-aa8d-4457-a4f9-31812b8107d3","Type":"ContainerDied","Data":"993c53a048dccfdb8658894e792ceda74488c299bcef3ba041508c31ec4d5811"} Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.241305 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c97027f1-aa8d-4457-a4f9-31812b8107d3","Type":"ContainerDied","Data":"5eec6e1d99857c1f2aeae8ff830dd6f2e27d9f8a3daceb05de765fbe66118fc6"} Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.241317 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5eec6e1d99857c1f2aeae8ff830dd6f2e27d9f8a3daceb05de765fbe66118fc6" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.244277 4904 generic.go:334] "Generic (PLEG): container finished" podID="17d3d79f-6c72-4ad5-862e-7a1e020333ae" containerID="8f05e84f5e134afd3c686d05171e1ea3c6c237b098109d8dd1f767f57a1224af" exitCode=0 Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.244313 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"17d3d79f-6c72-4ad5-862e-7a1e020333ae","Type":"ContainerDied","Data":"8f05e84f5e134afd3c686d05171e1ea3c6c237b098109d8dd1f767f57a1224af"} Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.244367 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"17d3d79f-6c72-4ad5-862e-7a1e020333ae","Type":"ContainerDied","Data":"ad0b20c3b709bd40b259990c265a99d903db27b5bc60f59d4b16dad628f52288"} Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.244369 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.244430 4904 scope.go:117] "RemoveContainer" containerID="8f05e84f5e134afd3c686d05171e1ea3c6c237b098109d8dd1f767f57a1224af" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.244869 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.260105 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17d3d79f-6c72-4ad5-862e-7a1e020333ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17d3d79f-6c72-4ad5-862e-7a1e020333ae" (UID: "17d3d79f-6c72-4ad5-862e-7a1e020333ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.274193 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17d3d79f-6c72-4ad5-862e-7a1e020333ae-config-data" (OuterVolumeSpecName: "config-data") pod "17d3d79f-6c72-4ad5-862e-7a1e020333ae" (UID: "17d3d79f-6c72-4ad5-862e-7a1e020333ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.274237 4904 scope.go:117] "RemoveContainer" containerID="5345ff80e0c415277499d8c6fda6bae4aafcadad8c75412238ce591e76bd0ea6" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.312106 4904 scope.go:117] "RemoveContainer" containerID="8f05e84f5e134afd3c686d05171e1ea3c6c237b098109d8dd1f767f57a1224af" Feb 14 11:29:19 crc kubenswrapper[4904]: E0214 11:29:19.312432 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f05e84f5e134afd3c686d05171e1ea3c6c237b098109d8dd1f767f57a1224af\": container with ID starting with 8f05e84f5e134afd3c686d05171e1ea3c6c237b098109d8dd1f767f57a1224af not found: ID does not exist" containerID="8f05e84f5e134afd3c686d05171e1ea3c6c237b098109d8dd1f767f57a1224af" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.312463 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f05e84f5e134afd3c686d05171e1ea3c6c237b098109d8dd1f767f57a1224af"} err="failed to get container status \"8f05e84f5e134afd3c686d05171e1ea3c6c237b098109d8dd1f767f57a1224af\": rpc error: code = NotFound desc = could not find container \"8f05e84f5e134afd3c686d05171e1ea3c6c237b098109d8dd1f767f57a1224af\": container with ID starting with 8f05e84f5e134afd3c686d05171e1ea3c6c237b098109d8dd1f767f57a1224af not found: ID does not exist" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.312483 4904 scope.go:117] "RemoveContainer" containerID="5345ff80e0c415277499d8c6fda6bae4aafcadad8c75412238ce591e76bd0ea6" Feb 14 11:29:19 crc kubenswrapper[4904]: E0214 11:29:19.313446 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5345ff80e0c415277499d8c6fda6bae4aafcadad8c75412238ce591e76bd0ea6\": container with ID starting with 5345ff80e0c415277499d8c6fda6bae4aafcadad8c75412238ce591e76bd0ea6 not found: ID does not exist" containerID="5345ff80e0c415277499d8c6fda6bae4aafcadad8c75412238ce591e76bd0ea6" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.313474 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5345ff80e0c415277499d8c6fda6bae4aafcadad8c75412238ce591e76bd0ea6"} err="failed to get container status \"5345ff80e0c415277499d8c6fda6bae4aafcadad8c75412238ce591e76bd0ea6\": rpc error: code = NotFound desc = could not find container \"5345ff80e0c415277499d8c6fda6bae4aafcadad8c75412238ce591e76bd0ea6\": container with ID starting with 5345ff80e0c415277499d8c6fda6bae4aafcadad8c75412238ce591e76bd0ea6 not found: ID does not exist" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.321346 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c97027f1-aa8d-4457-a4f9-31812b8107d3-config-data\") pod \"c97027f1-aa8d-4457-a4f9-31812b8107d3\" (UID: \"c97027f1-aa8d-4457-a4f9-31812b8107d3\") " Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.321514 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjk56\" (UniqueName: \"kubernetes.io/projected/c97027f1-aa8d-4457-a4f9-31812b8107d3-kube-api-access-mjk56\") pod \"c97027f1-aa8d-4457-a4f9-31812b8107d3\" (UID: \"c97027f1-aa8d-4457-a4f9-31812b8107d3\") " Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.321562 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c97027f1-aa8d-4457-a4f9-31812b8107d3-combined-ca-bundle\") pod \"c97027f1-aa8d-4457-a4f9-31812b8107d3\" (UID: \"c97027f1-aa8d-4457-a4f9-31812b8107d3\") " Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.321946 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17d3d79f-6c72-4ad5-862e-7a1e020333ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.321966 4904 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17d3d79f-6c72-4ad5-862e-7a1e020333ae-logs\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.321975 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17d3d79f-6c72-4ad5-862e-7a1e020333ae-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.321983 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nwlg\" (UniqueName: \"kubernetes.io/projected/17d3d79f-6c72-4ad5-862e-7a1e020333ae-kube-api-access-5nwlg\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.328128 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c97027f1-aa8d-4457-a4f9-31812b8107d3-kube-api-access-mjk56" (OuterVolumeSpecName: "kube-api-access-mjk56") pod "c97027f1-aa8d-4457-a4f9-31812b8107d3" (UID: "c97027f1-aa8d-4457-a4f9-31812b8107d3"). InnerVolumeSpecName "kube-api-access-mjk56". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.353384 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c97027f1-aa8d-4457-a4f9-31812b8107d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c97027f1-aa8d-4457-a4f9-31812b8107d3" (UID: "c97027f1-aa8d-4457-a4f9-31812b8107d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.360263 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c97027f1-aa8d-4457-a4f9-31812b8107d3-config-data" (OuterVolumeSpecName: "config-data") pod "c97027f1-aa8d-4457-a4f9-31812b8107d3" (UID: "c97027f1-aa8d-4457-a4f9-31812b8107d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.423943 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjk56\" (UniqueName: \"kubernetes.io/projected/c97027f1-aa8d-4457-a4f9-31812b8107d3-kube-api-access-mjk56\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.424191 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c97027f1-aa8d-4457-a4f9-31812b8107d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.424329 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c97027f1-aa8d-4457-a4f9-31812b8107d3-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.571710 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.580541 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.605456 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 14 11:29:19 crc kubenswrapper[4904]: E0214 11:29:19.606110 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c97027f1-aa8d-4457-a4f9-31812b8107d3" containerName="nova-scheduler-scheduler" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.606185 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="c97027f1-aa8d-4457-a4f9-31812b8107d3" containerName="nova-scheduler-scheduler" Feb 14 11:29:19 crc kubenswrapper[4904]: E0214 11:29:19.606266 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d3d79f-6c72-4ad5-862e-7a1e020333ae" containerName="nova-api-log" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.606329 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d3d79f-6c72-4ad5-862e-7a1e020333ae" containerName="nova-api-log" Feb 14 11:29:19 crc kubenswrapper[4904]: E0214 11:29:19.606389 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d3d79f-6c72-4ad5-862e-7a1e020333ae" containerName="nova-api-api" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.608469 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d3d79f-6c72-4ad5-862e-7a1e020333ae" containerName="nova-api-api" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.608863 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="17d3d79f-6c72-4ad5-862e-7a1e020333ae" containerName="nova-api-api" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.608946 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="17d3d79f-6c72-4ad5-862e-7a1e020333ae" containerName="nova-api-log" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.609018 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="c97027f1-aa8d-4457-a4f9-31812b8107d3" containerName="nova-scheduler-scheduler" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.609999 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.613221 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.617148 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.628752 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srsqg\" (UniqueName: \"kubernetes.io/projected/60bccbea-43ae-40fc-91a9-acf30bf4f114-kube-api-access-srsqg\") pod \"nova-api-0\" (UID: \"60bccbea-43ae-40fc-91a9-acf30bf4f114\") " pod="openstack/nova-api-0" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.628823 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60bccbea-43ae-40fc-91a9-acf30bf4f114-logs\") pod \"nova-api-0\" (UID: \"60bccbea-43ae-40fc-91a9-acf30bf4f114\") " pod="openstack/nova-api-0" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.628941 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60bccbea-43ae-40fc-91a9-acf30bf4f114-config-data\") pod \"nova-api-0\" (UID: \"60bccbea-43ae-40fc-91a9-acf30bf4f114\") " pod="openstack/nova-api-0" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.628964 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60bccbea-43ae-40fc-91a9-acf30bf4f114-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"60bccbea-43ae-40fc-91a9-acf30bf4f114\") " pod="openstack/nova-api-0" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.730861 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60bccbea-43ae-40fc-91a9-acf30bf4f114-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"60bccbea-43ae-40fc-91a9-acf30bf4f114\") " pod="openstack/nova-api-0" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.730902 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60bccbea-43ae-40fc-91a9-acf30bf4f114-config-data\") pod \"nova-api-0\" (UID: \"60bccbea-43ae-40fc-91a9-acf30bf4f114\") " pod="openstack/nova-api-0" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.730975 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srsqg\" (UniqueName: \"kubernetes.io/projected/60bccbea-43ae-40fc-91a9-acf30bf4f114-kube-api-access-srsqg\") pod \"nova-api-0\" (UID: \"60bccbea-43ae-40fc-91a9-acf30bf4f114\") " pod="openstack/nova-api-0" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.731003 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60bccbea-43ae-40fc-91a9-acf30bf4f114-logs\") pod \"nova-api-0\" (UID: \"60bccbea-43ae-40fc-91a9-acf30bf4f114\") " pod="openstack/nova-api-0" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.731754 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60bccbea-43ae-40fc-91a9-acf30bf4f114-logs\") pod \"nova-api-0\" (UID: \"60bccbea-43ae-40fc-91a9-acf30bf4f114\") " pod="openstack/nova-api-0" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.734667 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60bccbea-43ae-40fc-91a9-acf30bf4f114-config-data\") pod \"nova-api-0\" (UID: \"60bccbea-43ae-40fc-91a9-acf30bf4f114\") " pod="openstack/nova-api-0" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.735141 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60bccbea-43ae-40fc-91a9-acf30bf4f114-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"60bccbea-43ae-40fc-91a9-acf30bf4f114\") " pod="openstack/nova-api-0" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.746746 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srsqg\" (UniqueName: \"kubernetes.io/projected/60bccbea-43ae-40fc-91a9-acf30bf4f114-kube-api-access-srsqg\") pod \"nova-api-0\" (UID: \"60bccbea-43ae-40fc-91a9-acf30bf4f114\") " pod="openstack/nova-api-0" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.859200 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17d3d79f-6c72-4ad5-862e-7a1e020333ae" path="/var/lib/kubelet/pods/17d3d79f-6c72-4ad5-862e-7a1e020333ae/volumes" Feb 14 11:29:19 crc kubenswrapper[4904]: I0214 11:29:19.936465 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.254230 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.284321 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.297754 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.305983 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.307129 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.313530 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.337791 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.343253 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/395d1b56-d8ac-4d0c-9b35-17950f8e1da3-config-data\") pod \"nova-scheduler-0\" (UID: \"395d1b56-d8ac-4d0c-9b35-17950f8e1da3\") " pod="openstack/nova-scheduler-0" Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.343436 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzpjd\" (UniqueName: \"kubernetes.io/projected/395d1b56-d8ac-4d0c-9b35-17950f8e1da3-kube-api-access-pzpjd\") pod \"nova-scheduler-0\" (UID: \"395d1b56-d8ac-4d0c-9b35-17950f8e1da3\") " pod="openstack/nova-scheduler-0" Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.343508 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/395d1b56-d8ac-4d0c-9b35-17950f8e1da3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"395d1b56-d8ac-4d0c-9b35-17950f8e1da3\") " pod="openstack/nova-scheduler-0" Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.444757 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzpjd\" (UniqueName: \"kubernetes.io/projected/395d1b56-d8ac-4d0c-9b35-17950f8e1da3-kube-api-access-pzpjd\") pod \"nova-scheduler-0\" (UID: \"395d1b56-d8ac-4d0c-9b35-17950f8e1da3\") " pod="openstack/nova-scheduler-0" Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.445014 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/395d1b56-d8ac-4d0c-9b35-17950f8e1da3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"395d1b56-d8ac-4d0c-9b35-17950f8e1da3\") " pod="openstack/nova-scheduler-0" Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.445167 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/395d1b56-d8ac-4d0c-9b35-17950f8e1da3-config-data\") pod \"nova-scheduler-0\" (UID: \"395d1b56-d8ac-4d0c-9b35-17950f8e1da3\") " pod="openstack/nova-scheduler-0" Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.451046 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/395d1b56-d8ac-4d0c-9b35-17950f8e1da3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"395d1b56-d8ac-4d0c-9b35-17950f8e1da3\") " pod="openstack/nova-scheduler-0" Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.463855 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/395d1b56-d8ac-4d0c-9b35-17950f8e1da3-config-data\") pod \"nova-scheduler-0\" (UID: \"395d1b56-d8ac-4d0c-9b35-17950f8e1da3\") " pod="openstack/nova-scheduler-0" Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.481647 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzpjd\" (UniqueName: \"kubernetes.io/projected/395d1b56-d8ac-4d0c-9b35-17950f8e1da3-kube-api-access-pzpjd\") pod \"nova-scheduler-0\" (UID: \"395d1b56-d8ac-4d0c-9b35-17950f8e1da3\") " pod="openstack/nova-scheduler-0" Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.502393 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.636537 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.836349 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 14 11:29:20 crc kubenswrapper[4904]: I0214 11:29:20.837493 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 14 11:29:21 crc kubenswrapper[4904]: I0214 11:29:21.119427 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 14 11:29:21 crc kubenswrapper[4904]: I0214 11:29:21.264438 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"395d1b56-d8ac-4d0c-9b35-17950f8e1da3","Type":"ContainerStarted","Data":"1c8fcef3391f1b88c02d3a7be07e4d8e79d14814f406e84cc66b77590bf81bdb"} Feb 14 11:29:21 crc kubenswrapper[4904]: I0214 11:29:21.271260 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60bccbea-43ae-40fc-91a9-acf30bf4f114","Type":"ContainerStarted","Data":"747cdbac9f0aecf62342bd9ac946e4ae8eb7421b0f2f641644f25b7ff2e4c7a8"} Feb 14 11:29:21 crc kubenswrapper[4904]: I0214 11:29:21.271316 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60bccbea-43ae-40fc-91a9-acf30bf4f114","Type":"ContainerStarted","Data":"12337f7e0d034f949002167ff8741dc5d5669ab4d1a5e83d69527eec2836135a"} Feb 14 11:29:21 crc kubenswrapper[4904]: I0214 11:29:21.271341 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60bccbea-43ae-40fc-91a9-acf30bf4f114","Type":"ContainerStarted","Data":"bb074a8f99757ee7ed25f45324a8c740b08c70196cd23b6093b29767ef43fcb4"} Feb 14 11:29:21 crc kubenswrapper[4904]: I0214 11:29:21.296104 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.2960879739999998 podStartE2EDuration="2.296087974s" podCreationTimestamp="2026-02-14 11:29:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:29:21.288209317 +0000 UTC m=+1152.100973978" watchObservedRunningTime="2026-02-14 11:29:21.296087974 +0000 UTC m=+1152.108852635" Feb 14 11:29:21 crc kubenswrapper[4904]: I0214 11:29:21.846605 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c97027f1-aa8d-4457-a4f9-31812b8107d3" path="/var/lib/kubelet/pods/c97027f1-aa8d-4457-a4f9-31812b8107d3/volumes" Feb 14 11:29:22 crc kubenswrapper[4904]: I0214 11:29:22.280412 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"395d1b56-d8ac-4d0c-9b35-17950f8e1da3","Type":"ContainerStarted","Data":"4bb40494658f13cf27aa4e454f8af6af19851d098b00fa02d313310db4794fb9"} Feb 14 11:29:22 crc kubenswrapper[4904]: I0214 11:29:22.303607 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.303592158 podStartE2EDuration="2.303592158s" podCreationTimestamp="2026-02-14 11:29:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:29:22.299422415 +0000 UTC m=+1153.112187086" watchObservedRunningTime="2026-02-14 11:29:22.303592158 +0000 UTC m=+1153.116356819" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.080701 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-687657b496-tz596" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.213859 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-horizon-secret-key\") pod \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.214318 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-logs\") pod \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.214397 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-config-data\") pod \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.214448 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-scripts\") pod \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.214504 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-horizon-tls-certs\") pod \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.214598 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-combined-ca-bundle\") pod \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.214632 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x79gg\" (UniqueName: \"kubernetes.io/projected/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-kube-api-access-x79gg\") pod \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\" (UID: \"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393\") " Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.215094 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-logs" (OuterVolumeSpecName: "logs") pod "5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" (UID: "5d9d3948-4a4d-4d10-a6b6-d6d35eba3393"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.222802 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" (UID: "5d9d3948-4a4d-4d10-a6b6-d6d35eba3393"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.227958 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-kube-api-access-x79gg" (OuterVolumeSpecName: "kube-api-access-x79gg") pod "5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" (UID: "5d9d3948-4a4d-4d10-a6b6-d6d35eba3393"). InnerVolumeSpecName "kube-api-access-x79gg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.255650 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" (UID: "5d9d3948-4a4d-4d10-a6b6-d6d35eba3393"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.267592 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-config-data" (OuterVolumeSpecName: "config-data") pod "5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" (UID: "5d9d3948-4a4d-4d10-a6b6-d6d35eba3393"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.279298 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" (UID: "5d9d3948-4a4d-4d10-a6b6-d6d35eba3393"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.283013 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-scripts" (OuterVolumeSpecName: "scripts") pod "5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" (UID: "5d9d3948-4a4d-4d10-a6b6-d6d35eba3393"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.317485 4904 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.317750 4904 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-logs\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.317820 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.317923 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.317991 4904 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.318055 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.318122 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x79gg\" (UniqueName: \"kubernetes.io/projected/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393-kube-api-access-x79gg\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.318542 4904 generic.go:334] "Generic (PLEG): container finished" podID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerID="582bc45e7c749f481aab958e03a53b45ff101180b75c9d8a70ed2568b4a2948b" exitCode=137 Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.318582 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-687657b496-tz596" event={"ID":"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393","Type":"ContainerDied","Data":"582bc45e7c749f481aab958e03a53b45ff101180b75c9d8a70ed2568b4a2948b"} Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.318609 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-687657b496-tz596" event={"ID":"5d9d3948-4a4d-4d10-a6b6-d6d35eba3393","Type":"ContainerDied","Data":"b8527c01a195ffb25243d7944cf0993c62d9b22869ca946577bf35f1a1a95a82"} Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.318626 4904 scope.go:117] "RemoveContainer" containerID="a6c1837a2344350a20faa0c4c53cf29edd5c3937922bb9a13dd8aa63719d0470" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.318771 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-687657b496-tz596" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.358944 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-687657b496-tz596"] Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.365218 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-687657b496-tz596"] Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.492416 4904 scope.go:117] "RemoveContainer" containerID="582bc45e7c749f481aab958e03a53b45ff101180b75c9d8a70ed2568b4a2948b" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.514051 4904 scope.go:117] "RemoveContainer" containerID="a6c1837a2344350a20faa0c4c53cf29edd5c3937922bb9a13dd8aa63719d0470" Feb 14 11:29:24 crc kubenswrapper[4904]: E0214 11:29:24.515222 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6c1837a2344350a20faa0c4c53cf29edd5c3937922bb9a13dd8aa63719d0470\": container with ID starting with a6c1837a2344350a20faa0c4c53cf29edd5c3937922bb9a13dd8aa63719d0470 not found: ID does not exist" containerID="a6c1837a2344350a20faa0c4c53cf29edd5c3937922bb9a13dd8aa63719d0470" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.515261 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6c1837a2344350a20faa0c4c53cf29edd5c3937922bb9a13dd8aa63719d0470"} err="failed to get container status \"a6c1837a2344350a20faa0c4c53cf29edd5c3937922bb9a13dd8aa63719d0470\": rpc error: code = NotFound desc = could not find container \"a6c1837a2344350a20faa0c4c53cf29edd5c3937922bb9a13dd8aa63719d0470\": container with ID starting with a6c1837a2344350a20faa0c4c53cf29edd5c3937922bb9a13dd8aa63719d0470 not found: ID does not exist" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.515287 4904 scope.go:117] "RemoveContainer" containerID="582bc45e7c749f481aab958e03a53b45ff101180b75c9d8a70ed2568b4a2948b" Feb 14 11:29:24 crc kubenswrapper[4904]: E0214 11:29:24.515627 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"582bc45e7c749f481aab958e03a53b45ff101180b75c9d8a70ed2568b4a2948b\": container with ID starting with 582bc45e7c749f481aab958e03a53b45ff101180b75c9d8a70ed2568b4a2948b not found: ID does not exist" containerID="582bc45e7c749f481aab958e03a53b45ff101180b75c9d8a70ed2568b4a2948b" Feb 14 11:29:24 crc kubenswrapper[4904]: I0214 11:29:24.515668 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"582bc45e7c749f481aab958e03a53b45ff101180b75c9d8a70ed2568b4a2948b"} err="failed to get container status \"582bc45e7c749f481aab958e03a53b45ff101180b75c9d8a70ed2568b4a2948b\": rpc error: code = NotFound desc = could not find container \"582bc45e7c749f481aab958e03a53b45ff101180b75c9d8a70ed2568b4a2948b\": container with ID starting with 582bc45e7c749f481aab958e03a53b45ff101180b75c9d8a70ed2568b4a2948b not found: ID does not exist" Feb 14 11:29:25 crc kubenswrapper[4904]: I0214 11:29:25.538991 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 14 11:29:25 crc kubenswrapper[4904]: I0214 11:29:25.637681 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 14 11:29:25 crc kubenswrapper[4904]: I0214 11:29:25.825989 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 14 11:29:25 crc kubenswrapper[4904]: I0214 11:29:25.826035 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 14 11:29:25 crc kubenswrapper[4904]: I0214 11:29:25.846800 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" path="/var/lib/kubelet/pods/5d9d3948-4a4d-4d10-a6b6-d6d35eba3393/volumes" Feb 14 11:29:26 crc kubenswrapper[4904]: I0214 11:29:26.839989 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5b248689-1264-4fac-a77e-60f6494317f3" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 14 11:29:26 crc kubenswrapper[4904]: I0214 11:29:26.840103 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5b248689-1264-4fac-a77e-60f6494317f3" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 14 11:29:29 crc kubenswrapper[4904]: I0214 11:29:29.403232 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 14 11:29:29 crc kubenswrapper[4904]: I0214 11:29:29.936813 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 14 11:29:29 crc kubenswrapper[4904]: I0214 11:29:29.937133 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 14 11:29:30 crc kubenswrapper[4904]: I0214 11:29:30.637885 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 14 11:29:30 crc kubenswrapper[4904]: I0214 11:29:30.667544 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 14 11:29:31 crc kubenswrapper[4904]: I0214 11:29:31.019984 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="60bccbea-43ae-40fc-91a9-acf30bf4f114" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.199:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 14 11:29:31 crc kubenswrapper[4904]: I0214 11:29:31.020262 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="60bccbea-43ae-40fc-91a9-acf30bf4f114" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.199:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 14 11:29:31 crc kubenswrapper[4904]: I0214 11:29:31.431160 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 14 11:29:33 crc kubenswrapper[4904]: I0214 11:29:33.000513 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 14 11:29:33 crc kubenswrapper[4904]: I0214 11:29:33.001057 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="8515da2f-15c7-43af-b57f-bec14a6b0398" containerName="kube-state-metrics" containerID="cri-o://cabf8e56e34a6df0b7dde1428f2b80975722c3e1dc55210f459dd51cd7ab17a7" gracePeriod=30 Feb 14 11:29:33 crc kubenswrapper[4904]: I0214 11:29:33.423744 4904 generic.go:334] "Generic (PLEG): container finished" podID="8515da2f-15c7-43af-b57f-bec14a6b0398" containerID="cabf8e56e34a6df0b7dde1428f2b80975722c3e1dc55210f459dd51cd7ab17a7" exitCode=2 Feb 14 11:29:33 crc kubenswrapper[4904]: I0214 11:29:33.423920 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8515da2f-15c7-43af-b57f-bec14a6b0398","Type":"ContainerDied","Data":"cabf8e56e34a6df0b7dde1428f2b80975722c3e1dc55210f459dd51cd7ab17a7"} Feb 14 11:29:33 crc kubenswrapper[4904]: I0214 11:29:33.424072 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8515da2f-15c7-43af-b57f-bec14a6b0398","Type":"ContainerDied","Data":"c98bd77a5d4126fefa6d1970fdac9858507a41c462445f4c117339d578519602"} Feb 14 11:29:33 crc kubenswrapper[4904]: I0214 11:29:33.424086 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c98bd77a5d4126fefa6d1970fdac9858507a41c462445f4c117339d578519602" Feb 14 11:29:33 crc kubenswrapper[4904]: I0214 11:29:33.496192 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 14 11:29:33 crc kubenswrapper[4904]: I0214 11:29:33.580900 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pssqp\" (UniqueName: \"kubernetes.io/projected/8515da2f-15c7-43af-b57f-bec14a6b0398-kube-api-access-pssqp\") pod \"8515da2f-15c7-43af-b57f-bec14a6b0398\" (UID: \"8515da2f-15c7-43af-b57f-bec14a6b0398\") " Feb 14 11:29:33 crc kubenswrapper[4904]: I0214 11:29:33.589045 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8515da2f-15c7-43af-b57f-bec14a6b0398-kube-api-access-pssqp" (OuterVolumeSpecName: "kube-api-access-pssqp") pod "8515da2f-15c7-43af-b57f-bec14a6b0398" (UID: "8515da2f-15c7-43af-b57f-bec14a6b0398"). InnerVolumeSpecName "kube-api-access-pssqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:29:33 crc kubenswrapper[4904]: I0214 11:29:33.683547 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pssqp\" (UniqueName: \"kubernetes.io/projected/8515da2f-15c7-43af-b57f-bec14a6b0398-kube-api-access-pssqp\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.432509 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.455022 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.465005 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.515142 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 14 11:29:34 crc kubenswrapper[4904]: E0214 11:29:34.515541 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon-log" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.515557 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon-log" Feb 14 11:29:34 crc kubenswrapper[4904]: E0214 11:29:34.515578 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.515584 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon" Feb 14 11:29:34 crc kubenswrapper[4904]: E0214 11:29:34.515597 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8515da2f-15c7-43af-b57f-bec14a6b0398" containerName="kube-state-metrics" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.515603 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="8515da2f-15c7-43af-b57f-bec14a6b0398" containerName="kube-state-metrics" Feb 14 11:29:34 crc kubenswrapper[4904]: E0214 11:29:34.515614 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.515619 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.515798 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.515823 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="8515da2f-15c7-43af-b57f-bec14a6b0398" containerName="kube-state-metrics" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.515848 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon-log" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.516464 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.520598 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.520674 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.526320 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.602548 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5bdd6b9-9f22-4613-8536-99fdd632d2dd-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"e5bdd6b9-9f22-4613-8536-99fdd632d2dd\") " pod="openstack/kube-state-metrics-0" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.602583 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5bdd6b9-9f22-4613-8536-99fdd632d2dd-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"e5bdd6b9-9f22-4613-8536-99fdd632d2dd\") " pod="openstack/kube-state-metrics-0" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.602628 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/e5bdd6b9-9f22-4613-8536-99fdd632d2dd-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"e5bdd6b9-9f22-4613-8536-99fdd632d2dd\") " pod="openstack/kube-state-metrics-0" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.602657 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd5jv\" (UniqueName: \"kubernetes.io/projected/e5bdd6b9-9f22-4613-8536-99fdd632d2dd-kube-api-access-xd5jv\") pod \"kube-state-metrics-0\" (UID: \"e5bdd6b9-9f22-4613-8536-99fdd632d2dd\") " pod="openstack/kube-state-metrics-0" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.704273 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd5jv\" (UniqueName: \"kubernetes.io/projected/e5bdd6b9-9f22-4613-8536-99fdd632d2dd-kube-api-access-xd5jv\") pod \"kube-state-metrics-0\" (UID: \"e5bdd6b9-9f22-4613-8536-99fdd632d2dd\") " pod="openstack/kube-state-metrics-0" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.704422 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5bdd6b9-9f22-4613-8536-99fdd632d2dd-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"e5bdd6b9-9f22-4613-8536-99fdd632d2dd\") " pod="openstack/kube-state-metrics-0" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.704447 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5bdd6b9-9f22-4613-8536-99fdd632d2dd-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"e5bdd6b9-9f22-4613-8536-99fdd632d2dd\") " pod="openstack/kube-state-metrics-0" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.704487 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/e5bdd6b9-9f22-4613-8536-99fdd632d2dd-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"e5bdd6b9-9f22-4613-8536-99fdd632d2dd\") " pod="openstack/kube-state-metrics-0" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.708686 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/e5bdd6b9-9f22-4613-8536-99fdd632d2dd-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"e5bdd6b9-9f22-4613-8536-99fdd632d2dd\") " pod="openstack/kube-state-metrics-0" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.710537 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5bdd6b9-9f22-4613-8536-99fdd632d2dd-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"e5bdd6b9-9f22-4613-8536-99fdd632d2dd\") " pod="openstack/kube-state-metrics-0" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.712508 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5bdd6b9-9f22-4613-8536-99fdd632d2dd-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"e5bdd6b9-9f22-4613-8536-99fdd632d2dd\") " pod="openstack/kube-state-metrics-0" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.726144 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd5jv\" (UniqueName: \"kubernetes.io/projected/e5bdd6b9-9f22-4613-8536-99fdd632d2dd-kube-api-access-xd5jv\") pod \"kube-state-metrics-0\" (UID: \"e5bdd6b9-9f22-4613-8536-99fdd632d2dd\") " pod="openstack/kube-state-metrics-0" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.835536 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.973919 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.974577 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3169126-e313-4d31-8984-34e321aa0bde" containerName="ceilometer-central-agent" containerID="cri-o://5a83e64acd6041c6fc9c5b408baa8ff73b6d313aa58936932d5b78aa7c9c9fba" gracePeriod=30 Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.975163 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3169126-e313-4d31-8984-34e321aa0bde" containerName="proxy-httpd" containerID="cri-o://61e971e371cdfdc3d16eb2223dd330107dd8f408c1ad5d7a954aae8ae2346fdb" gracePeriod=30 Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.976426 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3169126-e313-4d31-8984-34e321aa0bde" containerName="ceilometer-notification-agent" containerID="cri-o://1d5ed9aad79c96147a4d24724ac5f8e70224f7b6336846171a4811621b800cc1" gracePeriod=30 Feb 14 11:29:34 crc kubenswrapper[4904]: I0214 11:29:34.976556 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3169126-e313-4d31-8984-34e321aa0bde" containerName="sg-core" containerID="cri-o://60a52b88d53fc54434a37fb8d8c53c5875d31f5b8923fb2285756021aa4e4c18" gracePeriod=30 Feb 14 11:29:35 crc kubenswrapper[4904]: I0214 11:29:35.369343 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 14 11:29:35 crc kubenswrapper[4904]: W0214 11:29:35.373733 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5bdd6b9_9f22_4613_8536_99fdd632d2dd.slice/crio-9c473e0b1fd243d8b5f20a587e45357747f65cac46098a02149b23a095c3ff28 WatchSource:0}: Error finding container 9c473e0b1fd243d8b5f20a587e45357747f65cac46098a02149b23a095c3ff28: Status 404 returned error can't find the container with id 9c473e0b1fd243d8b5f20a587e45357747f65cac46098a02149b23a095c3ff28 Feb 14 11:29:35 crc kubenswrapper[4904]: I0214 11:29:35.377065 4904 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 14 11:29:35 crc kubenswrapper[4904]: I0214 11:29:35.443436 4904 generic.go:334] "Generic (PLEG): container finished" podID="e3169126-e313-4d31-8984-34e321aa0bde" containerID="61e971e371cdfdc3d16eb2223dd330107dd8f408c1ad5d7a954aae8ae2346fdb" exitCode=0 Feb 14 11:29:35 crc kubenswrapper[4904]: I0214 11:29:35.444471 4904 generic.go:334] "Generic (PLEG): container finished" podID="e3169126-e313-4d31-8984-34e321aa0bde" containerID="60a52b88d53fc54434a37fb8d8c53c5875d31f5b8923fb2285756021aa4e4c18" exitCode=2 Feb 14 11:29:35 crc kubenswrapper[4904]: I0214 11:29:35.444585 4904 generic.go:334] "Generic (PLEG): container finished" podID="e3169126-e313-4d31-8984-34e321aa0bde" containerID="5a83e64acd6041c6fc9c5b408baa8ff73b6d313aa58936932d5b78aa7c9c9fba" exitCode=0 Feb 14 11:29:35 crc kubenswrapper[4904]: I0214 11:29:35.443511 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3169126-e313-4d31-8984-34e321aa0bde","Type":"ContainerDied","Data":"61e971e371cdfdc3d16eb2223dd330107dd8f408c1ad5d7a954aae8ae2346fdb"} Feb 14 11:29:35 crc kubenswrapper[4904]: I0214 11:29:35.444779 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3169126-e313-4d31-8984-34e321aa0bde","Type":"ContainerDied","Data":"60a52b88d53fc54434a37fb8d8c53c5875d31f5b8923fb2285756021aa4e4c18"} Feb 14 11:29:35 crc kubenswrapper[4904]: I0214 11:29:35.444892 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3169126-e313-4d31-8984-34e321aa0bde","Type":"ContainerDied","Data":"5a83e64acd6041c6fc9c5b408baa8ff73b6d313aa58936932d5b78aa7c9c9fba"} Feb 14 11:29:35 crc kubenswrapper[4904]: I0214 11:29:35.446186 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e5bdd6b9-9f22-4613-8536-99fdd632d2dd","Type":"ContainerStarted","Data":"9c473e0b1fd243d8b5f20a587e45357747f65cac46098a02149b23a095c3ff28"} Feb 14 11:29:35 crc kubenswrapper[4904]: I0214 11:29:35.846097 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8515da2f-15c7-43af-b57f-bec14a6b0398" path="/var/lib/kubelet/pods/8515da2f-15c7-43af-b57f-bec14a6b0398/volumes" Feb 14 11:29:35 crc kubenswrapper[4904]: I0214 11:29:35.846666 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 14 11:29:35 crc kubenswrapper[4904]: I0214 11:29:35.846714 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 14 11:29:35 crc kubenswrapper[4904]: I0214 11:29:35.852673 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 14 11:29:35 crc kubenswrapper[4904]: I0214 11:29:35.860187 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.086149 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.233874 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3169126-e313-4d31-8984-34e321aa0bde-log-httpd\") pod \"e3169126-e313-4d31-8984-34e321aa0bde\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.233975 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-scripts\") pod \"e3169126-e313-4d31-8984-34e321aa0bde\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.234031 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shrjv\" (UniqueName: \"kubernetes.io/projected/e3169126-e313-4d31-8984-34e321aa0bde-kube-api-access-shrjv\") pod \"e3169126-e313-4d31-8984-34e321aa0bde\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.234134 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3169126-e313-4d31-8984-34e321aa0bde-run-httpd\") pod \"e3169126-e313-4d31-8984-34e321aa0bde\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.234485 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3169126-e313-4d31-8984-34e321aa0bde-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e3169126-e313-4d31-8984-34e321aa0bde" (UID: "e3169126-e313-4d31-8984-34e321aa0bde"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.234572 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-combined-ca-bundle\") pod \"e3169126-e313-4d31-8984-34e321aa0bde\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.234656 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-config-data\") pod \"e3169126-e313-4d31-8984-34e321aa0bde\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.234696 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-sg-core-conf-yaml\") pod \"e3169126-e313-4d31-8984-34e321aa0bde\" (UID: \"e3169126-e313-4d31-8984-34e321aa0bde\") " Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.235631 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3169126-e313-4d31-8984-34e321aa0bde-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e3169126-e313-4d31-8984-34e321aa0bde" (UID: "e3169126-e313-4d31-8984-34e321aa0bde"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.235967 4904 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3169126-e313-4d31-8984-34e321aa0bde-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.235990 4904 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3169126-e313-4d31-8984-34e321aa0bde-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.239109 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-scripts" (OuterVolumeSpecName: "scripts") pod "e3169126-e313-4d31-8984-34e321aa0bde" (UID: "e3169126-e313-4d31-8984-34e321aa0bde"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.240911 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3169126-e313-4d31-8984-34e321aa0bde-kube-api-access-shrjv" (OuterVolumeSpecName: "kube-api-access-shrjv") pod "e3169126-e313-4d31-8984-34e321aa0bde" (UID: "e3169126-e313-4d31-8984-34e321aa0bde"). InnerVolumeSpecName "kube-api-access-shrjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.280616 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e3169126-e313-4d31-8984-34e321aa0bde" (UID: "e3169126-e313-4d31-8984-34e321aa0bde"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.318333 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3169126-e313-4d31-8984-34e321aa0bde" (UID: "e3169126-e313-4d31-8984-34e321aa0bde"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.337539 4904 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.337573 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.337585 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shrjv\" (UniqueName: \"kubernetes.io/projected/e3169126-e313-4d31-8984-34e321aa0bde-kube-api-access-shrjv\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.337595 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.341295 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-config-data" (OuterVolumeSpecName: "config-data") pod "e3169126-e313-4d31-8984-34e321aa0bde" (UID: "e3169126-e313-4d31-8984-34e321aa0bde"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.438918 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3169126-e313-4d31-8984-34e321aa0bde-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.457593 4904 generic.go:334] "Generic (PLEG): container finished" podID="e3169126-e313-4d31-8984-34e321aa0bde" containerID="1d5ed9aad79c96147a4d24724ac5f8e70224f7b6336846171a4811621b800cc1" exitCode=0 Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.457694 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3169126-e313-4d31-8984-34e321aa0bde","Type":"ContainerDied","Data":"1d5ed9aad79c96147a4d24724ac5f8e70224f7b6336846171a4811621b800cc1"} Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.457780 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3169126-e313-4d31-8984-34e321aa0bde","Type":"ContainerDied","Data":"8053e19b9a6494b14671b7fe2c963ab59d4bf1dee43b6803b3c1b1bbaed05490"} Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.457808 4904 scope.go:117] "RemoveContainer" containerID="61e971e371cdfdc3d16eb2223dd330107dd8f408c1ad5d7a954aae8ae2346fdb" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.458028 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.460022 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e5bdd6b9-9f22-4613-8536-99fdd632d2dd","Type":"ContainerStarted","Data":"529518d66b6474960d746308d47f7051d469b22571fe2c0411714fcdd553688c"} Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.475477 4904 scope.go:117] "RemoveContainer" containerID="60a52b88d53fc54434a37fb8d8c53c5875d31f5b8923fb2285756021aa4e4c18" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.498582 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.989344291 podStartE2EDuration="2.498560553s" podCreationTimestamp="2026-02-14 11:29:34 +0000 UTC" firstStartedPulling="2026-02-14 11:29:35.376743595 +0000 UTC m=+1166.189508256" lastFinishedPulling="2026-02-14 11:29:35.885959857 +0000 UTC m=+1166.698724518" observedRunningTime="2026-02-14 11:29:36.49298815 +0000 UTC m=+1167.305752811" watchObservedRunningTime="2026-02-14 11:29:36.498560553 +0000 UTC m=+1167.311325224" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.503197 4904 scope.go:117] "RemoveContainer" containerID="1d5ed9aad79c96147a4d24724ac5f8e70224f7b6336846171a4811621b800cc1" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.528288 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.529435 4904 scope.go:117] "RemoveContainer" containerID="5a83e64acd6041c6fc9c5b408baa8ff73b6d313aa58936932d5b78aa7c9c9fba" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.556444 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.564531 4904 scope.go:117] "RemoveContainer" containerID="61e971e371cdfdc3d16eb2223dd330107dd8f408c1ad5d7a954aae8ae2346fdb" Feb 14 11:29:36 crc kubenswrapper[4904]: E0214 11:29:36.566008 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61e971e371cdfdc3d16eb2223dd330107dd8f408c1ad5d7a954aae8ae2346fdb\": container with ID starting with 61e971e371cdfdc3d16eb2223dd330107dd8f408c1ad5d7a954aae8ae2346fdb not found: ID does not exist" containerID="61e971e371cdfdc3d16eb2223dd330107dd8f408c1ad5d7a954aae8ae2346fdb" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.566053 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61e971e371cdfdc3d16eb2223dd330107dd8f408c1ad5d7a954aae8ae2346fdb"} err="failed to get container status \"61e971e371cdfdc3d16eb2223dd330107dd8f408c1ad5d7a954aae8ae2346fdb\": rpc error: code = NotFound desc = could not find container \"61e971e371cdfdc3d16eb2223dd330107dd8f408c1ad5d7a954aae8ae2346fdb\": container with ID starting with 61e971e371cdfdc3d16eb2223dd330107dd8f408c1ad5d7a954aae8ae2346fdb not found: ID does not exist" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.566095 4904 scope.go:117] "RemoveContainer" containerID="60a52b88d53fc54434a37fb8d8c53c5875d31f5b8923fb2285756021aa4e4c18" Feb 14 11:29:36 crc kubenswrapper[4904]: E0214 11:29:36.566364 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60a52b88d53fc54434a37fb8d8c53c5875d31f5b8923fb2285756021aa4e4c18\": container with ID starting with 60a52b88d53fc54434a37fb8d8c53c5875d31f5b8923fb2285756021aa4e4c18 not found: ID does not exist" containerID="60a52b88d53fc54434a37fb8d8c53c5875d31f5b8923fb2285756021aa4e4c18" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.566389 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60a52b88d53fc54434a37fb8d8c53c5875d31f5b8923fb2285756021aa4e4c18"} err="failed to get container status \"60a52b88d53fc54434a37fb8d8c53c5875d31f5b8923fb2285756021aa4e4c18\": rpc error: code = NotFound desc = could not find container \"60a52b88d53fc54434a37fb8d8c53c5875d31f5b8923fb2285756021aa4e4c18\": container with ID starting with 60a52b88d53fc54434a37fb8d8c53c5875d31f5b8923fb2285756021aa4e4c18 not found: ID does not exist" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.566406 4904 scope.go:117] "RemoveContainer" containerID="1d5ed9aad79c96147a4d24724ac5f8e70224f7b6336846171a4811621b800cc1" Feb 14 11:29:36 crc kubenswrapper[4904]: E0214 11:29:36.566657 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d5ed9aad79c96147a4d24724ac5f8e70224f7b6336846171a4811621b800cc1\": container with ID starting with 1d5ed9aad79c96147a4d24724ac5f8e70224f7b6336846171a4811621b800cc1 not found: ID does not exist" containerID="1d5ed9aad79c96147a4d24724ac5f8e70224f7b6336846171a4811621b800cc1" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.566690 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d5ed9aad79c96147a4d24724ac5f8e70224f7b6336846171a4811621b800cc1"} err="failed to get container status \"1d5ed9aad79c96147a4d24724ac5f8e70224f7b6336846171a4811621b800cc1\": rpc error: code = NotFound desc = could not find container \"1d5ed9aad79c96147a4d24724ac5f8e70224f7b6336846171a4811621b800cc1\": container with ID starting with 1d5ed9aad79c96147a4d24724ac5f8e70224f7b6336846171a4811621b800cc1 not found: ID does not exist" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.566705 4904 scope.go:117] "RemoveContainer" containerID="5a83e64acd6041c6fc9c5b408baa8ff73b6d313aa58936932d5b78aa7c9c9fba" Feb 14 11:29:36 crc kubenswrapper[4904]: E0214 11:29:36.567135 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a83e64acd6041c6fc9c5b408baa8ff73b6d313aa58936932d5b78aa7c9c9fba\": container with ID starting with 5a83e64acd6041c6fc9c5b408baa8ff73b6d313aa58936932d5b78aa7c9c9fba not found: ID does not exist" containerID="5a83e64acd6041c6fc9c5b408baa8ff73b6d313aa58936932d5b78aa7c9c9fba" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.567165 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a83e64acd6041c6fc9c5b408baa8ff73b6d313aa58936932d5b78aa7c9c9fba"} err="failed to get container status \"5a83e64acd6041c6fc9c5b408baa8ff73b6d313aa58936932d5b78aa7c9c9fba\": rpc error: code = NotFound desc = could not find container \"5a83e64acd6041c6fc9c5b408baa8ff73b6d313aa58936932d5b78aa7c9c9fba\": container with ID starting with 5a83e64acd6041c6fc9c5b408baa8ff73b6d313aa58936932d5b78aa7c9c9fba not found: ID does not exist" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.573952 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:29:36 crc kubenswrapper[4904]: E0214 11:29:36.574388 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3169126-e313-4d31-8984-34e321aa0bde" containerName="sg-core" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.574407 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3169126-e313-4d31-8984-34e321aa0bde" containerName="sg-core" Feb 14 11:29:36 crc kubenswrapper[4904]: E0214 11:29:36.574420 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3169126-e313-4d31-8984-34e321aa0bde" containerName="ceilometer-central-agent" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.574428 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3169126-e313-4d31-8984-34e321aa0bde" containerName="ceilometer-central-agent" Feb 14 11:29:36 crc kubenswrapper[4904]: E0214 11:29:36.574448 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3169126-e313-4d31-8984-34e321aa0bde" containerName="ceilometer-notification-agent" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.574455 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3169126-e313-4d31-8984-34e321aa0bde" containerName="ceilometer-notification-agent" Feb 14 11:29:36 crc kubenswrapper[4904]: E0214 11:29:36.574463 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3169126-e313-4d31-8984-34e321aa0bde" containerName="proxy-httpd" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.574469 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3169126-e313-4d31-8984-34e321aa0bde" containerName="proxy-httpd" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.574647 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3169126-e313-4d31-8984-34e321aa0bde" containerName="ceilometer-central-agent" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.574664 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3169126-e313-4d31-8984-34e321aa0bde" containerName="proxy-httpd" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.574681 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d9d3948-4a4d-4d10-a6b6-d6d35eba3393" containerName="horizon" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.574689 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3169126-e313-4d31-8984-34e321aa0bde" containerName="ceilometer-notification-agent" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.574699 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3169126-e313-4d31-8984-34e321aa0bde" containerName="sg-core" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.576284 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.580190 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.580352 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.580464 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.584973 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.643788 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltwnn\" (UniqueName: \"kubernetes.io/projected/dece8216-82c6-466c-81b9-45a92bd3b5c3-kube-api-access-ltwnn\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.643874 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dece8216-82c6-466c-81b9-45a92bd3b5c3-log-httpd\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.644045 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-config-data\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.644207 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.644263 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-scripts\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.644306 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.644436 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.644518 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dece8216-82c6-466c-81b9-45a92bd3b5c3-run-httpd\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.746170 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-config-data\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.746255 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.746281 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-scripts\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.746310 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.746986 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.747783 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dece8216-82c6-466c-81b9-45a92bd3b5c3-run-httpd\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.748176 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dece8216-82c6-466c-81b9-45a92bd3b5c3-run-httpd\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.748369 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltwnn\" (UniqueName: \"kubernetes.io/projected/dece8216-82c6-466c-81b9-45a92bd3b5c3-kube-api-access-ltwnn\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.748698 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dece8216-82c6-466c-81b9-45a92bd3b5c3-log-httpd\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.749157 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dece8216-82c6-466c-81b9-45a92bd3b5c3-log-httpd\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.750665 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.753028 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.753437 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.753670 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-config-data\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.755529 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-scripts\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.773780 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltwnn\" (UniqueName: \"kubernetes.io/projected/dece8216-82c6-466c-81b9-45a92bd3b5c3-kube-api-access-ltwnn\") pod \"ceilometer-0\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " pod="openstack/ceilometer-0" Feb 14 11:29:36 crc kubenswrapper[4904]: I0214 11:29:36.907478 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:29:37 crc kubenswrapper[4904]: I0214 11:29:37.478540 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 14 11:29:37 crc kubenswrapper[4904]: W0214 11:29:37.544297 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddece8216_82c6_466c_81b9_45a92bd3b5c3.slice/crio-d0b472bf94c7e6d1b67b5d4dcd12f8d101214371ec99bc092a9819f26557fa1b WatchSource:0}: Error finding container d0b472bf94c7e6d1b67b5d4dcd12f8d101214371ec99bc092a9819f26557fa1b: Status 404 returned error can't find the container with id d0b472bf94c7e6d1b67b5d4dcd12f8d101214371ec99bc092a9819f26557fa1b Feb 14 11:29:37 crc kubenswrapper[4904]: I0214 11:29:37.546950 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:29:37 crc kubenswrapper[4904]: I0214 11:29:37.851388 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3169126-e313-4d31-8984-34e321aa0bde" path="/var/lib/kubelet/pods/e3169126-e313-4d31-8984-34e321aa0bde/volumes" Feb 14 11:29:38 crc kubenswrapper[4904]: E0214 11:29:38.132274 4904 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod905037ac_d5f7_46f8_b0ce_aa2649771221.slice/crio-5e6f50efb3246fc04bd274dc7fa75e9387dd28c5ea36f158cf11a23757d93c10.scope\": RecentStats: unable to find data in memory cache]" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.342286 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.377541 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/905037ac-d5f7-46f8-b0ce-aa2649771221-config-data\") pod \"905037ac-d5f7-46f8-b0ce-aa2649771221\" (UID: \"905037ac-d5f7-46f8-b0ce-aa2649771221\") " Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.377614 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/905037ac-d5f7-46f8-b0ce-aa2649771221-combined-ca-bundle\") pod \"905037ac-d5f7-46f8-b0ce-aa2649771221\" (UID: \"905037ac-d5f7-46f8-b0ce-aa2649771221\") " Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.377705 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwsmg\" (UniqueName: \"kubernetes.io/projected/905037ac-d5f7-46f8-b0ce-aa2649771221-kube-api-access-bwsmg\") pod \"905037ac-d5f7-46f8-b0ce-aa2649771221\" (UID: \"905037ac-d5f7-46f8-b0ce-aa2649771221\") " Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.389151 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/905037ac-d5f7-46f8-b0ce-aa2649771221-kube-api-access-bwsmg" (OuterVolumeSpecName: "kube-api-access-bwsmg") pod "905037ac-d5f7-46f8-b0ce-aa2649771221" (UID: "905037ac-d5f7-46f8-b0ce-aa2649771221"). InnerVolumeSpecName "kube-api-access-bwsmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.404470 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/905037ac-d5f7-46f8-b0ce-aa2649771221-config-data" (OuterVolumeSpecName: "config-data") pod "905037ac-d5f7-46f8-b0ce-aa2649771221" (UID: "905037ac-d5f7-46f8-b0ce-aa2649771221"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.412546 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/905037ac-d5f7-46f8-b0ce-aa2649771221-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "905037ac-d5f7-46f8-b0ce-aa2649771221" (UID: "905037ac-d5f7-46f8-b0ce-aa2649771221"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.479435 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/905037ac-d5f7-46f8-b0ce-aa2649771221-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.479464 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/905037ac-d5f7-46f8-b0ce-aa2649771221-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.479476 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwsmg\" (UniqueName: \"kubernetes.io/projected/905037ac-d5f7-46f8-b0ce-aa2649771221-kube-api-access-bwsmg\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.486004 4904 generic.go:334] "Generic (PLEG): container finished" podID="905037ac-d5f7-46f8-b0ce-aa2649771221" containerID="5e6f50efb3246fc04bd274dc7fa75e9387dd28c5ea36f158cf11a23757d93c10" exitCode=137 Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.486061 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.486079 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"905037ac-d5f7-46f8-b0ce-aa2649771221","Type":"ContainerDied","Data":"5e6f50efb3246fc04bd274dc7fa75e9387dd28c5ea36f158cf11a23757d93c10"} Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.486118 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"905037ac-d5f7-46f8-b0ce-aa2649771221","Type":"ContainerDied","Data":"dc40692325f5a3fb4b212d427cb4ea383d11d6cf7ce146d2b45187d8b80bf76b"} Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.486137 4904 scope.go:117] "RemoveContainer" containerID="5e6f50efb3246fc04bd274dc7fa75e9387dd28c5ea36f158cf11a23757d93c10" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.487663 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dece8216-82c6-466c-81b9-45a92bd3b5c3","Type":"ContainerStarted","Data":"492744c72c3faf071069fb1f2311ec1d7cff0e9bc2909b35abcbe64c85eac21d"} Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.487705 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dece8216-82c6-466c-81b9-45a92bd3b5c3","Type":"ContainerStarted","Data":"d0b472bf94c7e6d1b67b5d4dcd12f8d101214371ec99bc092a9819f26557fa1b"} Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.508147 4904 scope.go:117] "RemoveContainer" containerID="5e6f50efb3246fc04bd274dc7fa75e9387dd28c5ea36f158cf11a23757d93c10" Feb 14 11:29:38 crc kubenswrapper[4904]: E0214 11:29:38.508627 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e6f50efb3246fc04bd274dc7fa75e9387dd28c5ea36f158cf11a23757d93c10\": container with ID starting with 5e6f50efb3246fc04bd274dc7fa75e9387dd28c5ea36f158cf11a23757d93c10 not found: ID does not exist" containerID="5e6f50efb3246fc04bd274dc7fa75e9387dd28c5ea36f158cf11a23757d93c10" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.508654 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e6f50efb3246fc04bd274dc7fa75e9387dd28c5ea36f158cf11a23757d93c10"} err="failed to get container status \"5e6f50efb3246fc04bd274dc7fa75e9387dd28c5ea36f158cf11a23757d93c10\": rpc error: code = NotFound desc = could not find container \"5e6f50efb3246fc04bd274dc7fa75e9387dd28c5ea36f158cf11a23757d93c10\": container with ID starting with 5e6f50efb3246fc04bd274dc7fa75e9387dd28c5ea36f158cf11a23757d93c10 not found: ID does not exist" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.518993 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.527435 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.543766 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 14 11:29:38 crc kubenswrapper[4904]: E0214 11:29:38.544151 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="905037ac-d5f7-46f8-b0ce-aa2649771221" containerName="nova-cell1-novncproxy-novncproxy" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.544171 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="905037ac-d5f7-46f8-b0ce-aa2649771221" containerName="nova-cell1-novncproxy-novncproxy" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.544360 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="905037ac-d5f7-46f8-b0ce-aa2649771221" containerName="nova-cell1-novncproxy-novncproxy" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.544959 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.547892 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.548114 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.548292 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.569948 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.583691 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c89ca73a-9933-4fc8-a96e-48489f786747-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c89ca73a-9933-4fc8-a96e-48489f786747\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.583763 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-729mb\" (UniqueName: \"kubernetes.io/projected/c89ca73a-9933-4fc8-a96e-48489f786747-kube-api-access-729mb\") pod \"nova-cell1-novncproxy-0\" (UID: \"c89ca73a-9933-4fc8-a96e-48489f786747\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.583846 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c89ca73a-9933-4fc8-a96e-48489f786747-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c89ca73a-9933-4fc8-a96e-48489f786747\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.584046 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c89ca73a-9933-4fc8-a96e-48489f786747-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c89ca73a-9933-4fc8-a96e-48489f786747\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.584195 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c89ca73a-9933-4fc8-a96e-48489f786747-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c89ca73a-9933-4fc8-a96e-48489f786747\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.685744 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c89ca73a-9933-4fc8-a96e-48489f786747-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c89ca73a-9933-4fc8-a96e-48489f786747\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.685830 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c89ca73a-9933-4fc8-a96e-48489f786747-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c89ca73a-9933-4fc8-a96e-48489f786747\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.685900 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c89ca73a-9933-4fc8-a96e-48489f786747-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c89ca73a-9933-4fc8-a96e-48489f786747\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.685923 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-729mb\" (UniqueName: \"kubernetes.io/projected/c89ca73a-9933-4fc8-a96e-48489f786747-kube-api-access-729mb\") pod \"nova-cell1-novncproxy-0\" (UID: \"c89ca73a-9933-4fc8-a96e-48489f786747\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.685958 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c89ca73a-9933-4fc8-a96e-48489f786747-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c89ca73a-9933-4fc8-a96e-48489f786747\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.689905 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c89ca73a-9933-4fc8-a96e-48489f786747-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c89ca73a-9933-4fc8-a96e-48489f786747\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.690763 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c89ca73a-9933-4fc8-a96e-48489f786747-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c89ca73a-9933-4fc8-a96e-48489f786747\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.691816 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c89ca73a-9933-4fc8-a96e-48489f786747-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c89ca73a-9933-4fc8-a96e-48489f786747\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.696208 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c89ca73a-9933-4fc8-a96e-48489f786747-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c89ca73a-9933-4fc8-a96e-48489f786747\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.704167 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-729mb\" (UniqueName: \"kubernetes.io/projected/c89ca73a-9933-4fc8-a96e-48489f786747-kube-api-access-729mb\") pod \"nova-cell1-novncproxy-0\" (UID: \"c89ca73a-9933-4fc8-a96e-48489f786747\") " pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:38 crc kubenswrapper[4904]: I0214 11:29:38.867466 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:39 crc kubenswrapper[4904]: I0214 11:29:39.331254 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 14 11:29:39 crc kubenswrapper[4904]: W0214 11:29:39.338468 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc89ca73a_9933_4fc8_a96e_48489f786747.slice/crio-e59c174598b2d86103d380c035048dec599c8adc188625c77f97bfaebe7a36e2 WatchSource:0}: Error finding container e59c174598b2d86103d380c035048dec599c8adc188625c77f97bfaebe7a36e2: Status 404 returned error can't find the container with id e59c174598b2d86103d380c035048dec599c8adc188625c77f97bfaebe7a36e2 Feb 14 11:29:39 crc kubenswrapper[4904]: I0214 11:29:39.498232 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dece8216-82c6-466c-81b9-45a92bd3b5c3","Type":"ContainerStarted","Data":"d4685cd3c3e2d5f10b5a2f6ab1d8c353b991c6bef9096872dc3666b840c34714"} Feb 14 11:29:39 crc kubenswrapper[4904]: I0214 11:29:39.498287 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dece8216-82c6-466c-81b9-45a92bd3b5c3","Type":"ContainerStarted","Data":"9a9be08b3280ebe4f1613a14c139cada904bfc4055109a9dd672d1c45bf5a795"} Feb 14 11:29:39 crc kubenswrapper[4904]: I0214 11:29:39.501650 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c89ca73a-9933-4fc8-a96e-48489f786747","Type":"ContainerStarted","Data":"e59c174598b2d86103d380c035048dec599c8adc188625c77f97bfaebe7a36e2"} Feb 14 11:29:39 crc kubenswrapper[4904]: I0214 11:29:39.848904 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="905037ac-d5f7-46f8-b0ce-aa2649771221" path="/var/lib/kubelet/pods/905037ac-d5f7-46f8-b0ce-aa2649771221/volumes" Feb 14 11:29:39 crc kubenswrapper[4904]: I0214 11:29:39.949327 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 14 11:29:39 crc kubenswrapper[4904]: I0214 11:29:39.952873 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 14 11:29:39 crc kubenswrapper[4904]: I0214 11:29:39.953406 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 14 11:29:39 crc kubenswrapper[4904]: I0214 11:29:39.962120 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.530217 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c89ca73a-9933-4fc8-a96e-48489f786747","Type":"ContainerStarted","Data":"2fd0cb66653957be4b2ebad0fd4140a20fef89e31c9a2e4b4e71eba9bd3a009c"} Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.530815 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.539916 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.556516 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.556498584 podStartE2EDuration="2.556498584s" podCreationTimestamp="2026-02-14 11:29:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:29:40.551882158 +0000 UTC m=+1171.364646829" watchObservedRunningTime="2026-02-14 11:29:40.556498584 +0000 UTC m=+1171.369263235" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.760374 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-bl2fg"] Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.762410 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.802139 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-bl2fg"] Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.850150 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-config\") pod \"dnsmasq-dns-89c5cd4d5-bl2fg\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.850203 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-bl2fg\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.850228 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-bl2fg\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.850255 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-bl2fg\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.850278 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxjfk\" (UniqueName: \"kubernetes.io/projected/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-kube-api-access-rxjfk\") pod \"dnsmasq-dns-89c5cd4d5-bl2fg\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.850323 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-bl2fg\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.952369 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-config\") pod \"dnsmasq-dns-89c5cd4d5-bl2fg\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.952444 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-bl2fg\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.952469 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-bl2fg\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.952502 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-bl2fg\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.952528 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxjfk\" (UniqueName: \"kubernetes.io/projected/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-kube-api-access-rxjfk\") pod \"dnsmasq-dns-89c5cd4d5-bl2fg\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.952593 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-bl2fg\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.953568 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-config\") pod \"dnsmasq-dns-89c5cd4d5-bl2fg\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.955073 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-bl2fg\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.955153 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-bl2fg\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.957561 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-bl2fg\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.958227 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-bl2fg\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:40 crc kubenswrapper[4904]: I0214 11:29:40.973260 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxjfk\" (UniqueName: \"kubernetes.io/projected/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-kube-api-access-rxjfk\") pod \"dnsmasq-dns-89c5cd4d5-bl2fg\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:41 crc kubenswrapper[4904]: I0214 11:29:41.101097 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:41 crc kubenswrapper[4904]: I0214 11:29:41.545564 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dece8216-82c6-466c-81b9-45a92bd3b5c3","Type":"ContainerStarted","Data":"b4de7ac7f307eb185d67ba48d6b190f856b6a9239aa9f1b1e1f0621bc242c17b"} Feb 14 11:29:41 crc kubenswrapper[4904]: I0214 11:29:41.546081 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 14 11:29:41 crc kubenswrapper[4904]: I0214 11:29:41.580715 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.68446821 podStartE2EDuration="5.580697558s" podCreationTimestamp="2026-02-14 11:29:36 +0000 UTC" firstStartedPulling="2026-02-14 11:29:37.546417147 +0000 UTC m=+1168.359181808" lastFinishedPulling="2026-02-14 11:29:40.442646495 +0000 UTC m=+1171.255411156" observedRunningTime="2026-02-14 11:29:41.573796939 +0000 UTC m=+1172.386561610" watchObservedRunningTime="2026-02-14 11:29:41.580697558 +0000 UTC m=+1172.393462219" Feb 14 11:29:41 crc kubenswrapper[4904]: I0214 11:29:41.717630 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-bl2fg"] Feb 14 11:29:42 crc kubenswrapper[4904]: I0214 11:29:42.556633 4904 generic.go:334] "Generic (PLEG): container finished" podID="0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41" containerID="f93e1d601a859b953290f628d69fdf5b1e1f08ec46d991cff6a9d8e5e65ca6b4" exitCode=0 Feb 14 11:29:42 crc kubenswrapper[4904]: I0214 11:29:42.556729 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" event={"ID":"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41","Type":"ContainerDied","Data":"f93e1d601a859b953290f628d69fdf5b1e1f08ec46d991cff6a9d8e5e65ca6b4"} Feb 14 11:29:42 crc kubenswrapper[4904]: I0214 11:29:42.556971 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" event={"ID":"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41","Type":"ContainerStarted","Data":"87f077076dde91c45574e80c731fffaa26eaeee68e929d83b7eb7f6ed139797d"} Feb 14 11:29:43 crc kubenswrapper[4904]: I0214 11:29:43.425027 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 14 11:29:43 crc kubenswrapper[4904]: I0214 11:29:43.566947 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="60bccbea-43ae-40fc-91a9-acf30bf4f114" containerName="nova-api-log" containerID="cri-o://12337f7e0d034f949002167ff8741dc5d5669ab4d1a5e83d69527eec2836135a" gracePeriod=30 Feb 14 11:29:43 crc kubenswrapper[4904]: I0214 11:29:43.567842 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" event={"ID":"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41","Type":"ContainerStarted","Data":"90052437023c3c0bd09b40b8d6216c637bcdc4e70e68e5a12b619632522d303c"} Feb 14 11:29:43 crc kubenswrapper[4904]: I0214 11:29:43.568430 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:43 crc kubenswrapper[4904]: I0214 11:29:43.568511 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="60bccbea-43ae-40fc-91a9-acf30bf4f114" containerName="nova-api-api" containerID="cri-o://747cdbac9f0aecf62342bd9ac946e4ae8eb7421b0f2f641644f25b7ff2e4c7a8" gracePeriod=30 Feb 14 11:29:43 crc kubenswrapper[4904]: I0214 11:29:43.867698 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:44 crc kubenswrapper[4904]: I0214 11:29:44.414238 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" podStartSLOduration=4.414216401 podStartE2EDuration="4.414216401s" podCreationTimestamp="2026-02-14 11:29:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:29:43.593273784 +0000 UTC m=+1174.406038455" watchObservedRunningTime="2026-02-14 11:29:44.414216401 +0000 UTC m=+1175.226981062" Feb 14 11:29:44 crc kubenswrapper[4904]: I0214 11:29:44.425098 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:29:44 crc kubenswrapper[4904]: I0214 11:29:44.576039 4904 generic.go:334] "Generic (PLEG): container finished" podID="60bccbea-43ae-40fc-91a9-acf30bf4f114" containerID="12337f7e0d034f949002167ff8741dc5d5669ab4d1a5e83d69527eec2836135a" exitCode=143 Feb 14 11:29:44 crc kubenswrapper[4904]: I0214 11:29:44.576306 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerName="ceilometer-central-agent" containerID="cri-o://492744c72c3faf071069fb1f2311ec1d7cff0e9bc2909b35abcbe64c85eac21d" gracePeriod=30 Feb 14 11:29:44 crc kubenswrapper[4904]: I0214 11:29:44.576378 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60bccbea-43ae-40fc-91a9-acf30bf4f114","Type":"ContainerDied","Data":"12337f7e0d034f949002167ff8741dc5d5669ab4d1a5e83d69527eec2836135a"} Feb 14 11:29:44 crc kubenswrapper[4904]: I0214 11:29:44.577173 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerName="proxy-httpd" containerID="cri-o://b4de7ac7f307eb185d67ba48d6b190f856b6a9239aa9f1b1e1f0621bc242c17b" gracePeriod=30 Feb 14 11:29:44 crc kubenswrapper[4904]: I0214 11:29:44.577231 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerName="sg-core" containerID="cri-o://d4685cd3c3e2d5f10b5a2f6ab1d8c353b991c6bef9096872dc3666b840c34714" gracePeriod=30 Feb 14 11:29:44 crc kubenswrapper[4904]: I0214 11:29:44.577266 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerName="ceilometer-notification-agent" containerID="cri-o://9a9be08b3280ebe4f1613a14c139cada904bfc4055109a9dd672d1c45bf5a795" gracePeriod=30 Feb 14 11:29:44 crc kubenswrapper[4904]: I0214 11:29:44.850260 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 14 11:29:45 crc kubenswrapper[4904]: I0214 11:29:45.599672 4904 generic.go:334] "Generic (PLEG): container finished" podID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerID="b4de7ac7f307eb185d67ba48d6b190f856b6a9239aa9f1b1e1f0621bc242c17b" exitCode=0 Feb 14 11:29:45 crc kubenswrapper[4904]: I0214 11:29:45.599705 4904 generic.go:334] "Generic (PLEG): container finished" podID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerID="d4685cd3c3e2d5f10b5a2f6ab1d8c353b991c6bef9096872dc3666b840c34714" exitCode=2 Feb 14 11:29:45 crc kubenswrapper[4904]: I0214 11:29:45.599714 4904 generic.go:334] "Generic (PLEG): container finished" podID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerID="9a9be08b3280ebe4f1613a14c139cada904bfc4055109a9dd672d1c45bf5a795" exitCode=0 Feb 14 11:29:45 crc kubenswrapper[4904]: I0214 11:29:45.599772 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dece8216-82c6-466c-81b9-45a92bd3b5c3","Type":"ContainerDied","Data":"b4de7ac7f307eb185d67ba48d6b190f856b6a9239aa9f1b1e1f0621bc242c17b"} Feb 14 11:29:45 crc kubenswrapper[4904]: I0214 11:29:45.599812 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dece8216-82c6-466c-81b9-45a92bd3b5c3","Type":"ContainerDied","Data":"d4685cd3c3e2d5f10b5a2f6ab1d8c353b991c6bef9096872dc3666b840c34714"} Feb 14 11:29:45 crc kubenswrapper[4904]: I0214 11:29:45.599827 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dece8216-82c6-466c-81b9-45a92bd3b5c3","Type":"ContainerDied","Data":"9a9be08b3280ebe4f1613a14c139cada904bfc4055109a9dd672d1c45bf5a795"} Feb 14 11:29:46 crc kubenswrapper[4904]: I0214 11:29:46.382830 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:29:46 crc kubenswrapper[4904]: I0214 11:29:46.383270 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:29:46 crc kubenswrapper[4904]: I0214 11:29:46.383336 4904 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:29:46 crc kubenswrapper[4904]: I0214 11:29:46.384428 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d567bf42576edbcc9fce256f34faf5a41303db84f6e9f300d9f0592cbb6a48e2"} pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 14 11:29:46 crc kubenswrapper[4904]: I0214 11:29:46.384532 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" containerID="cri-o://d567bf42576edbcc9fce256f34faf5a41303db84f6e9f300d9f0592cbb6a48e2" gracePeriod=600 Feb 14 11:29:46 crc kubenswrapper[4904]: I0214 11:29:46.610910 4904 generic.go:334] "Generic (PLEG): container finished" podID="7790d276-2506-4641-8c22-67e6441417f0" containerID="d567bf42576edbcc9fce256f34faf5a41303db84f6e9f300d9f0592cbb6a48e2" exitCode=0 Feb 14 11:29:46 crc kubenswrapper[4904]: I0214 11:29:46.610953 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerDied","Data":"d567bf42576edbcc9fce256f34faf5a41303db84f6e9f300d9f0592cbb6a48e2"} Feb 14 11:29:46 crc kubenswrapper[4904]: I0214 11:29:46.610984 4904 scope.go:117] "RemoveContainer" containerID="fd7271c1e14d135193bc38a4ef3eb1646cd74f3540f5fa874f75aef55d03613d" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.137621 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.186660 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60bccbea-43ae-40fc-91a9-acf30bf4f114-config-data\") pod \"60bccbea-43ae-40fc-91a9-acf30bf4f114\" (UID: \"60bccbea-43ae-40fc-91a9-acf30bf4f114\") " Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.186703 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60bccbea-43ae-40fc-91a9-acf30bf4f114-logs\") pod \"60bccbea-43ae-40fc-91a9-acf30bf4f114\" (UID: \"60bccbea-43ae-40fc-91a9-acf30bf4f114\") " Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.186822 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60bccbea-43ae-40fc-91a9-acf30bf4f114-combined-ca-bundle\") pod \"60bccbea-43ae-40fc-91a9-acf30bf4f114\" (UID: \"60bccbea-43ae-40fc-91a9-acf30bf4f114\") " Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.186898 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srsqg\" (UniqueName: \"kubernetes.io/projected/60bccbea-43ae-40fc-91a9-acf30bf4f114-kube-api-access-srsqg\") pod \"60bccbea-43ae-40fc-91a9-acf30bf4f114\" (UID: \"60bccbea-43ae-40fc-91a9-acf30bf4f114\") " Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.188258 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60bccbea-43ae-40fc-91a9-acf30bf4f114-logs" (OuterVolumeSpecName: "logs") pod "60bccbea-43ae-40fc-91a9-acf30bf4f114" (UID: "60bccbea-43ae-40fc-91a9-acf30bf4f114"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.193110 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60bccbea-43ae-40fc-91a9-acf30bf4f114-kube-api-access-srsqg" (OuterVolumeSpecName: "kube-api-access-srsqg") pod "60bccbea-43ae-40fc-91a9-acf30bf4f114" (UID: "60bccbea-43ae-40fc-91a9-acf30bf4f114"). InnerVolumeSpecName "kube-api-access-srsqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.237278 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60bccbea-43ae-40fc-91a9-acf30bf4f114-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60bccbea-43ae-40fc-91a9-acf30bf4f114" (UID: "60bccbea-43ae-40fc-91a9-acf30bf4f114"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.249616 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60bccbea-43ae-40fc-91a9-acf30bf4f114-config-data" (OuterVolumeSpecName: "config-data") pod "60bccbea-43ae-40fc-91a9-acf30bf4f114" (UID: "60bccbea-43ae-40fc-91a9-acf30bf4f114"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.289071 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srsqg\" (UniqueName: \"kubernetes.io/projected/60bccbea-43ae-40fc-91a9-acf30bf4f114-kube-api-access-srsqg\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.289105 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60bccbea-43ae-40fc-91a9-acf30bf4f114-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.289114 4904 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60bccbea-43ae-40fc-91a9-acf30bf4f114-logs\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.289123 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60bccbea-43ae-40fc-91a9-acf30bf4f114-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.624676 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"7753982cea734b21f061129ef8b3fab0b493e6bfec1d88e26788a0cb6fd15b70"} Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.632197 4904 generic.go:334] "Generic (PLEG): container finished" podID="60bccbea-43ae-40fc-91a9-acf30bf4f114" containerID="747cdbac9f0aecf62342bd9ac946e4ae8eb7421b0f2f641644f25b7ff2e4c7a8" exitCode=0 Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.632347 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60bccbea-43ae-40fc-91a9-acf30bf4f114","Type":"ContainerDied","Data":"747cdbac9f0aecf62342bd9ac946e4ae8eb7421b0f2f641644f25b7ff2e4c7a8"} Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.632469 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60bccbea-43ae-40fc-91a9-acf30bf4f114","Type":"ContainerDied","Data":"bb074a8f99757ee7ed25f45324a8c740b08c70196cd23b6093b29767ef43fcb4"} Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.632544 4904 scope.go:117] "RemoveContainer" containerID="747cdbac9f0aecf62342bd9ac946e4ae8eb7421b0f2f641644f25b7ff2e4c7a8" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.632415 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.673467 4904 scope.go:117] "RemoveContainer" containerID="12337f7e0d034f949002167ff8741dc5d5669ab4d1a5e83d69527eec2836135a" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.714713 4904 scope.go:117] "RemoveContainer" containerID="747cdbac9f0aecf62342bd9ac946e4ae8eb7421b0f2f641644f25b7ff2e4c7a8" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.714851 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 14 11:29:47 crc kubenswrapper[4904]: E0214 11:29:47.717104 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"747cdbac9f0aecf62342bd9ac946e4ae8eb7421b0f2f641644f25b7ff2e4c7a8\": container with ID starting with 747cdbac9f0aecf62342bd9ac946e4ae8eb7421b0f2f641644f25b7ff2e4c7a8 not found: ID does not exist" containerID="747cdbac9f0aecf62342bd9ac946e4ae8eb7421b0f2f641644f25b7ff2e4c7a8" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.717152 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"747cdbac9f0aecf62342bd9ac946e4ae8eb7421b0f2f641644f25b7ff2e4c7a8"} err="failed to get container status \"747cdbac9f0aecf62342bd9ac946e4ae8eb7421b0f2f641644f25b7ff2e4c7a8\": rpc error: code = NotFound desc = could not find container \"747cdbac9f0aecf62342bd9ac946e4ae8eb7421b0f2f641644f25b7ff2e4c7a8\": container with ID starting with 747cdbac9f0aecf62342bd9ac946e4ae8eb7421b0f2f641644f25b7ff2e4c7a8 not found: ID does not exist" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.717178 4904 scope.go:117] "RemoveContainer" containerID="12337f7e0d034f949002167ff8741dc5d5669ab4d1a5e83d69527eec2836135a" Feb 14 11:29:47 crc kubenswrapper[4904]: E0214 11:29:47.720227 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12337f7e0d034f949002167ff8741dc5d5669ab4d1a5e83d69527eec2836135a\": container with ID starting with 12337f7e0d034f949002167ff8741dc5d5669ab4d1a5e83d69527eec2836135a not found: ID does not exist" containerID="12337f7e0d034f949002167ff8741dc5d5669ab4d1a5e83d69527eec2836135a" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.720252 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12337f7e0d034f949002167ff8741dc5d5669ab4d1a5e83d69527eec2836135a"} err="failed to get container status \"12337f7e0d034f949002167ff8741dc5d5669ab4d1a5e83d69527eec2836135a\": rpc error: code = NotFound desc = could not find container \"12337f7e0d034f949002167ff8741dc5d5669ab4d1a5e83d69527eec2836135a\": container with ID starting with 12337f7e0d034f949002167ff8741dc5d5669ab4d1a5e83d69527eec2836135a not found: ID does not exist" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.725917 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.743806 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 14 11:29:47 crc kubenswrapper[4904]: E0214 11:29:47.744405 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60bccbea-43ae-40fc-91a9-acf30bf4f114" containerName="nova-api-api" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.744427 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="60bccbea-43ae-40fc-91a9-acf30bf4f114" containerName="nova-api-api" Feb 14 11:29:47 crc kubenswrapper[4904]: E0214 11:29:47.744440 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60bccbea-43ae-40fc-91a9-acf30bf4f114" containerName="nova-api-log" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.744447 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="60bccbea-43ae-40fc-91a9-acf30bf4f114" containerName="nova-api-log" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.744657 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="60bccbea-43ae-40fc-91a9-acf30bf4f114" containerName="nova-api-log" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.744687 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="60bccbea-43ae-40fc-91a9-acf30bf4f114" containerName="nova-api-api" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.745928 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.747935 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.748211 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.750453 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.763535 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.798640 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de9dd864-0953-44cf-ad93-5284648538f2-logs\") pod \"nova-api-0\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.798751 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.798772 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.798818 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-public-tls-certs\") pod \"nova-api-0\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.798873 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khd78\" (UniqueName: \"kubernetes.io/projected/de9dd864-0953-44cf-ad93-5284648538f2-kube-api-access-khd78\") pod \"nova-api-0\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.798901 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-config-data\") pod \"nova-api-0\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.848238 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60bccbea-43ae-40fc-91a9-acf30bf4f114" path="/var/lib/kubelet/pods/60bccbea-43ae-40fc-91a9-acf30bf4f114/volumes" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.900306 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de9dd864-0953-44cf-ad93-5284648538f2-logs\") pod \"nova-api-0\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.900437 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.900463 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.900518 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-public-tls-certs\") pod \"nova-api-0\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.900561 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khd78\" (UniqueName: \"kubernetes.io/projected/de9dd864-0953-44cf-ad93-5284648538f2-kube-api-access-khd78\") pod \"nova-api-0\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.900599 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-config-data\") pod \"nova-api-0\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.900858 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de9dd864-0953-44cf-ad93-5284648538f2-logs\") pod \"nova-api-0\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.906441 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.909772 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-config-data\") pod \"nova-api-0\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.910329 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-public-tls-certs\") pod \"nova-api-0\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.910616 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " pod="openstack/nova-api-0" Feb 14 11:29:47 crc kubenswrapper[4904]: I0214 11:29:47.925256 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khd78\" (UniqueName: \"kubernetes.io/projected/de9dd864-0953-44cf-ad93-5284648538f2-kube-api-access-khd78\") pod \"nova-api-0\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " pod="openstack/nova-api-0" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.069719 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.412884 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.513291 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-config-data\") pod \"dece8216-82c6-466c-81b9-45a92bd3b5c3\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.513332 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-sg-core-conf-yaml\") pod \"dece8216-82c6-466c-81b9-45a92bd3b5c3\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.513409 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-ceilometer-tls-certs\") pod \"dece8216-82c6-466c-81b9-45a92bd3b5c3\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.513465 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-combined-ca-bundle\") pod \"dece8216-82c6-466c-81b9-45a92bd3b5c3\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.513560 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dece8216-82c6-466c-81b9-45a92bd3b5c3-run-httpd\") pod \"dece8216-82c6-466c-81b9-45a92bd3b5c3\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.513603 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltwnn\" (UniqueName: \"kubernetes.io/projected/dece8216-82c6-466c-81b9-45a92bd3b5c3-kube-api-access-ltwnn\") pod \"dece8216-82c6-466c-81b9-45a92bd3b5c3\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.513637 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dece8216-82c6-466c-81b9-45a92bd3b5c3-log-httpd\") pod \"dece8216-82c6-466c-81b9-45a92bd3b5c3\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.513675 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-scripts\") pod \"dece8216-82c6-466c-81b9-45a92bd3b5c3\" (UID: \"dece8216-82c6-466c-81b9-45a92bd3b5c3\") " Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.519151 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-scripts" (OuterVolumeSpecName: "scripts") pod "dece8216-82c6-466c-81b9-45a92bd3b5c3" (UID: "dece8216-82c6-466c-81b9-45a92bd3b5c3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.522200 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dece8216-82c6-466c-81b9-45a92bd3b5c3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "dece8216-82c6-466c-81b9-45a92bd3b5c3" (UID: "dece8216-82c6-466c-81b9-45a92bd3b5c3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.523384 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dece8216-82c6-466c-81b9-45a92bd3b5c3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "dece8216-82c6-466c-81b9-45a92bd3b5c3" (UID: "dece8216-82c6-466c-81b9-45a92bd3b5c3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.529164 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dece8216-82c6-466c-81b9-45a92bd3b5c3-kube-api-access-ltwnn" (OuterVolumeSpecName: "kube-api-access-ltwnn") pod "dece8216-82c6-466c-81b9-45a92bd3b5c3" (UID: "dece8216-82c6-466c-81b9-45a92bd3b5c3"). InnerVolumeSpecName "kube-api-access-ltwnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.589671 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "dece8216-82c6-466c-81b9-45a92bd3b5c3" (UID: "dece8216-82c6-466c-81b9-45a92bd3b5c3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.616645 4904 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.616687 4904 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dece8216-82c6-466c-81b9-45a92bd3b5c3-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.616701 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltwnn\" (UniqueName: \"kubernetes.io/projected/dece8216-82c6-466c-81b9-45a92bd3b5c3-kube-api-access-ltwnn\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.616714 4904 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dece8216-82c6-466c-81b9-45a92bd3b5c3-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.616723 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.622415 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 14 11:29:48 crc kubenswrapper[4904]: W0214 11:29:48.635454 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde9dd864_0953_44cf_ad93_5284648538f2.slice/crio-7d0b6b5f18867da24a2a2a9fe18f01918092ea279ea17231f2ea8e7891a5cb65 WatchSource:0}: Error finding container 7d0b6b5f18867da24a2a2a9fe18f01918092ea279ea17231f2ea8e7891a5cb65: Status 404 returned error can't find the container with id 7d0b6b5f18867da24a2a2a9fe18f01918092ea279ea17231f2ea8e7891a5cb65 Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.650491 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "dece8216-82c6-466c-81b9-45a92bd3b5c3" (UID: "dece8216-82c6-466c-81b9-45a92bd3b5c3"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.655040 4904 generic.go:334] "Generic (PLEG): container finished" podID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerID="492744c72c3faf071069fb1f2311ec1d7cff0e9bc2909b35abcbe64c85eac21d" exitCode=0 Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.655166 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.655301 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dece8216-82c6-466c-81b9-45a92bd3b5c3","Type":"ContainerDied","Data":"492744c72c3faf071069fb1f2311ec1d7cff0e9bc2909b35abcbe64c85eac21d"} Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.655411 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dece8216-82c6-466c-81b9-45a92bd3b5c3","Type":"ContainerDied","Data":"d0b472bf94c7e6d1b67b5d4dcd12f8d101214371ec99bc092a9819f26557fa1b"} Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.655504 4904 scope.go:117] "RemoveContainer" containerID="b4de7ac7f307eb185d67ba48d6b190f856b6a9239aa9f1b1e1f0621bc242c17b" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.688304 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dece8216-82c6-466c-81b9-45a92bd3b5c3" (UID: "dece8216-82c6-466c-81b9-45a92bd3b5c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.717095 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-config-data" (OuterVolumeSpecName: "config-data") pod "dece8216-82c6-466c-81b9-45a92bd3b5c3" (UID: "dece8216-82c6-466c-81b9-45a92bd3b5c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.718288 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.718311 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.718319 4904 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dece8216-82c6-466c-81b9-45a92bd3b5c3-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.754037 4904 scope.go:117] "RemoveContainer" containerID="d4685cd3c3e2d5f10b5a2f6ab1d8c353b991c6bef9096872dc3666b840c34714" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.771433 4904 scope.go:117] "RemoveContainer" containerID="9a9be08b3280ebe4f1613a14c139cada904bfc4055109a9dd672d1c45bf5a795" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.795710 4904 scope.go:117] "RemoveContainer" containerID="492744c72c3faf071069fb1f2311ec1d7cff0e9bc2909b35abcbe64c85eac21d" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.815111 4904 scope.go:117] "RemoveContainer" containerID="b4de7ac7f307eb185d67ba48d6b190f856b6a9239aa9f1b1e1f0621bc242c17b" Feb 14 11:29:48 crc kubenswrapper[4904]: E0214 11:29:48.815704 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4de7ac7f307eb185d67ba48d6b190f856b6a9239aa9f1b1e1f0621bc242c17b\": container with ID starting with b4de7ac7f307eb185d67ba48d6b190f856b6a9239aa9f1b1e1f0621bc242c17b not found: ID does not exist" containerID="b4de7ac7f307eb185d67ba48d6b190f856b6a9239aa9f1b1e1f0621bc242c17b" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.815801 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4de7ac7f307eb185d67ba48d6b190f856b6a9239aa9f1b1e1f0621bc242c17b"} err="failed to get container status \"b4de7ac7f307eb185d67ba48d6b190f856b6a9239aa9f1b1e1f0621bc242c17b\": rpc error: code = NotFound desc = could not find container \"b4de7ac7f307eb185d67ba48d6b190f856b6a9239aa9f1b1e1f0621bc242c17b\": container with ID starting with b4de7ac7f307eb185d67ba48d6b190f856b6a9239aa9f1b1e1f0621bc242c17b not found: ID does not exist" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.815933 4904 scope.go:117] "RemoveContainer" containerID="d4685cd3c3e2d5f10b5a2f6ab1d8c353b991c6bef9096872dc3666b840c34714" Feb 14 11:29:48 crc kubenswrapper[4904]: E0214 11:29:48.816388 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4685cd3c3e2d5f10b5a2f6ab1d8c353b991c6bef9096872dc3666b840c34714\": container with ID starting with d4685cd3c3e2d5f10b5a2f6ab1d8c353b991c6bef9096872dc3666b840c34714 not found: ID does not exist" containerID="d4685cd3c3e2d5f10b5a2f6ab1d8c353b991c6bef9096872dc3666b840c34714" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.816500 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4685cd3c3e2d5f10b5a2f6ab1d8c353b991c6bef9096872dc3666b840c34714"} err="failed to get container status \"d4685cd3c3e2d5f10b5a2f6ab1d8c353b991c6bef9096872dc3666b840c34714\": rpc error: code = NotFound desc = could not find container \"d4685cd3c3e2d5f10b5a2f6ab1d8c353b991c6bef9096872dc3666b840c34714\": container with ID starting with d4685cd3c3e2d5f10b5a2f6ab1d8c353b991c6bef9096872dc3666b840c34714 not found: ID does not exist" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.816572 4904 scope.go:117] "RemoveContainer" containerID="9a9be08b3280ebe4f1613a14c139cada904bfc4055109a9dd672d1c45bf5a795" Feb 14 11:29:48 crc kubenswrapper[4904]: E0214 11:29:48.817107 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a9be08b3280ebe4f1613a14c139cada904bfc4055109a9dd672d1c45bf5a795\": container with ID starting with 9a9be08b3280ebe4f1613a14c139cada904bfc4055109a9dd672d1c45bf5a795 not found: ID does not exist" containerID="9a9be08b3280ebe4f1613a14c139cada904bfc4055109a9dd672d1c45bf5a795" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.817145 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a9be08b3280ebe4f1613a14c139cada904bfc4055109a9dd672d1c45bf5a795"} err="failed to get container status \"9a9be08b3280ebe4f1613a14c139cada904bfc4055109a9dd672d1c45bf5a795\": rpc error: code = NotFound desc = could not find container \"9a9be08b3280ebe4f1613a14c139cada904bfc4055109a9dd672d1c45bf5a795\": container with ID starting with 9a9be08b3280ebe4f1613a14c139cada904bfc4055109a9dd672d1c45bf5a795 not found: ID does not exist" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.817171 4904 scope.go:117] "RemoveContainer" containerID="492744c72c3faf071069fb1f2311ec1d7cff0e9bc2909b35abcbe64c85eac21d" Feb 14 11:29:48 crc kubenswrapper[4904]: E0214 11:29:48.817499 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"492744c72c3faf071069fb1f2311ec1d7cff0e9bc2909b35abcbe64c85eac21d\": container with ID starting with 492744c72c3faf071069fb1f2311ec1d7cff0e9bc2909b35abcbe64c85eac21d not found: ID does not exist" containerID="492744c72c3faf071069fb1f2311ec1d7cff0e9bc2909b35abcbe64c85eac21d" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.817595 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"492744c72c3faf071069fb1f2311ec1d7cff0e9bc2909b35abcbe64c85eac21d"} err="failed to get container status \"492744c72c3faf071069fb1f2311ec1d7cff0e9bc2909b35abcbe64c85eac21d\": rpc error: code = NotFound desc = could not find container \"492744c72c3faf071069fb1f2311ec1d7cff0e9bc2909b35abcbe64c85eac21d\": container with ID starting with 492744c72c3faf071069fb1f2311ec1d7cff0e9bc2909b35abcbe64c85eac21d not found: ID does not exist" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.868104 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:48 crc kubenswrapper[4904]: I0214 11:29:48.888876 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.006861 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.018432 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.061375 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:29:49 crc kubenswrapper[4904]: E0214 11:29:49.061903 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerName="ceilometer-central-agent" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.061924 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerName="ceilometer-central-agent" Feb 14 11:29:49 crc kubenswrapper[4904]: E0214 11:29:49.061948 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerName="sg-core" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.061956 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerName="sg-core" Feb 14 11:29:49 crc kubenswrapper[4904]: E0214 11:29:49.061972 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerName="ceilometer-notification-agent" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.061979 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerName="ceilometer-notification-agent" Feb 14 11:29:49 crc kubenswrapper[4904]: E0214 11:29:49.062000 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerName="proxy-httpd" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.062006 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerName="proxy-httpd" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.062175 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerName="sg-core" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.062238 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerName="ceilometer-notification-agent" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.062258 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerName="ceilometer-central-agent" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.062270 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="dece8216-82c6-466c-81b9-45a92bd3b5c3" containerName="proxy-httpd" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.064164 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.067642 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.068035 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.068250 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.077287 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.124699 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbe346d4-5e03-40bc-9901-09a4ab6da61c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.124770 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbe346d4-5e03-40bc-9901-09a4ab6da61c-config-data\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.124977 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbe346d4-5e03-40bc-9901-09a4ab6da61c-scripts\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.125053 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlpmq\" (UniqueName: \"kubernetes.io/projected/cbe346d4-5e03-40bc-9901-09a4ab6da61c-kube-api-access-vlpmq\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.125074 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cbe346d4-5e03-40bc-9901-09a4ab6da61c-log-httpd\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.125783 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cbe346d4-5e03-40bc-9901-09a4ab6da61c-run-httpd\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.125847 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cbe346d4-5e03-40bc-9901-09a4ab6da61c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.125884 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbe346d4-5e03-40bc-9901-09a4ab6da61c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.236963 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbe346d4-5e03-40bc-9901-09a4ab6da61c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.237056 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbe346d4-5e03-40bc-9901-09a4ab6da61c-config-data\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.237084 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbe346d4-5e03-40bc-9901-09a4ab6da61c-scripts\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.237184 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlpmq\" (UniqueName: \"kubernetes.io/projected/cbe346d4-5e03-40bc-9901-09a4ab6da61c-kube-api-access-vlpmq\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.237205 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cbe346d4-5e03-40bc-9901-09a4ab6da61c-log-httpd\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.237313 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cbe346d4-5e03-40bc-9901-09a4ab6da61c-run-httpd\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.237387 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cbe346d4-5e03-40bc-9901-09a4ab6da61c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.237435 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbe346d4-5e03-40bc-9901-09a4ab6da61c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.242192 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbe346d4-5e03-40bc-9901-09a4ab6da61c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.244742 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbe346d4-5e03-40bc-9901-09a4ab6da61c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.246618 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cbe346d4-5e03-40bc-9901-09a4ab6da61c-log-httpd\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.247189 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cbe346d4-5e03-40bc-9901-09a4ab6da61c-run-httpd\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.251007 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cbe346d4-5e03-40bc-9901-09a4ab6da61c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.257577 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbe346d4-5e03-40bc-9901-09a4ab6da61c-scripts\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.262709 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbe346d4-5e03-40bc-9901-09a4ab6da61c-config-data\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.264308 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlpmq\" (UniqueName: \"kubernetes.io/projected/cbe346d4-5e03-40bc-9901-09a4ab6da61c-kube-api-access-vlpmq\") pod \"ceilometer-0\" (UID: \"cbe346d4-5e03-40bc-9901-09a4ab6da61c\") " pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.391052 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.667074 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de9dd864-0953-44cf-ad93-5284648538f2","Type":"ContainerStarted","Data":"7709ecf24b0951207660c1134dbe1ed2247dab2b32054996f1d80f996e404312"} Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.667417 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de9dd864-0953-44cf-ad93-5284648538f2","Type":"ContainerStarted","Data":"000e175ba2a360712075d7b0cd3d6fc23f990e1253cbfb4384baefc541503401"} Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.667434 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de9dd864-0953-44cf-ad93-5284648538f2","Type":"ContainerStarted","Data":"7d0b6b5f18867da24a2a2a9fe18f01918092ea279ea17231f2ea8e7891a5cb65"} Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.690104 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.690084872 podStartE2EDuration="2.690084872s" podCreationTimestamp="2026-02-14 11:29:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:29:49.684378385 +0000 UTC m=+1180.497143046" watchObservedRunningTime="2026-02-14 11:29:49.690084872 +0000 UTC m=+1180.502849533" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.699920 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.863238 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dece8216-82c6-466c-81b9-45a92bd3b5c3" path="/var/lib/kubelet/pods/dece8216-82c6-466c-81b9-45a92bd3b5c3/volumes" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.863978 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 14 11:29:49 crc kubenswrapper[4904]: W0214 11:29:49.865189 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbe346d4_5e03_40bc_9901_09a4ab6da61c.slice/crio-8f63a3da58b5b62b13cb8df48d84ba2424a71295d0c505d7cb74dc3a42eb1b3b WatchSource:0}: Error finding container 8f63a3da58b5b62b13cb8df48d84ba2424a71295d0c505d7cb74dc3a42eb1b3b: Status 404 returned error can't find the container with id 8f63a3da58b5b62b13cb8df48d84ba2424a71295d0c505d7cb74dc3a42eb1b3b Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.883566 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-4kbtk"] Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.885028 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4kbtk" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.887323 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.887638 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.913328 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-4kbtk"] Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.951880 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4kbtk\" (UID: \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\") " pod="openstack/nova-cell1-cell-mapping-4kbtk" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.951959 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-scripts\") pod \"nova-cell1-cell-mapping-4kbtk\" (UID: \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\") " pod="openstack/nova-cell1-cell-mapping-4kbtk" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.951990 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jvng\" (UniqueName: \"kubernetes.io/projected/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-kube-api-access-8jvng\") pod \"nova-cell1-cell-mapping-4kbtk\" (UID: \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\") " pod="openstack/nova-cell1-cell-mapping-4kbtk" Feb 14 11:29:49 crc kubenswrapper[4904]: I0214 11:29:49.952094 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-config-data\") pod \"nova-cell1-cell-mapping-4kbtk\" (UID: \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\") " pod="openstack/nova-cell1-cell-mapping-4kbtk" Feb 14 11:29:50 crc kubenswrapper[4904]: I0214 11:29:50.054320 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-config-data\") pod \"nova-cell1-cell-mapping-4kbtk\" (UID: \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\") " pod="openstack/nova-cell1-cell-mapping-4kbtk" Feb 14 11:29:50 crc kubenswrapper[4904]: I0214 11:29:50.054437 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4kbtk\" (UID: \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\") " pod="openstack/nova-cell1-cell-mapping-4kbtk" Feb 14 11:29:50 crc kubenswrapper[4904]: I0214 11:29:50.054494 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-scripts\") pod \"nova-cell1-cell-mapping-4kbtk\" (UID: \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\") " pod="openstack/nova-cell1-cell-mapping-4kbtk" Feb 14 11:29:50 crc kubenswrapper[4904]: I0214 11:29:50.054526 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jvng\" (UniqueName: \"kubernetes.io/projected/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-kube-api-access-8jvng\") pod \"nova-cell1-cell-mapping-4kbtk\" (UID: \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\") " pod="openstack/nova-cell1-cell-mapping-4kbtk" Feb 14 11:29:50 crc kubenswrapper[4904]: I0214 11:29:50.059960 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4kbtk\" (UID: \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\") " pod="openstack/nova-cell1-cell-mapping-4kbtk" Feb 14 11:29:50 crc kubenswrapper[4904]: I0214 11:29:50.060959 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-config-data\") pod \"nova-cell1-cell-mapping-4kbtk\" (UID: \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\") " pod="openstack/nova-cell1-cell-mapping-4kbtk" Feb 14 11:29:50 crc kubenswrapper[4904]: I0214 11:29:50.061225 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-scripts\") pod \"nova-cell1-cell-mapping-4kbtk\" (UID: \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\") " pod="openstack/nova-cell1-cell-mapping-4kbtk" Feb 14 11:29:50 crc kubenswrapper[4904]: I0214 11:29:50.071529 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jvng\" (UniqueName: \"kubernetes.io/projected/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-kube-api-access-8jvng\") pod \"nova-cell1-cell-mapping-4kbtk\" (UID: \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\") " pod="openstack/nova-cell1-cell-mapping-4kbtk" Feb 14 11:29:50 crc kubenswrapper[4904]: I0214 11:29:50.219549 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4kbtk" Feb 14 11:29:51 crc kubenswrapper[4904]: I0214 11:29:50.680617 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cbe346d4-5e03-40bc-9901-09a4ab6da61c","Type":"ContainerStarted","Data":"3f7bff010be11251fbfd0716c209879d9394b094ea677e760b9abad2cc2a3922"} Feb 14 11:29:51 crc kubenswrapper[4904]: I0214 11:29:50.680975 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cbe346d4-5e03-40bc-9901-09a4ab6da61c","Type":"ContainerStarted","Data":"8f63a3da58b5b62b13cb8df48d84ba2424a71295d0c505d7cb74dc3a42eb1b3b"} Feb 14 11:29:51 crc kubenswrapper[4904]: I0214 11:29:51.102955 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:29:51 crc kubenswrapper[4904]: I0214 11:29:51.163989 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-9qq9d"] Feb 14 11:29:51 crc kubenswrapper[4904]: I0214 11:29:51.164213 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" podUID="37dfca9c-dcd2-4620-b6c6-55225fc0693a" containerName="dnsmasq-dns" containerID="cri-o://8bbdd4200ba9f703d1be47f0df0a36df1548e458965b28c4ef87d2a9757f4ec7" gracePeriod=10 Feb 14 11:29:51 crc kubenswrapper[4904]: I0214 11:29:51.495887 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-4kbtk"] Feb 14 11:29:51 crc kubenswrapper[4904]: I0214 11:29:51.695554 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4kbtk" event={"ID":"29681a87-9c4c-47fd-9834-c3cf2ebd79d3","Type":"ContainerStarted","Data":"ae80c5c61032338d6856d97291ad28aac95f0950ebecae35e2308b3d9c164820"} Feb 14 11:29:51 crc kubenswrapper[4904]: I0214 11:29:51.698995 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cbe346d4-5e03-40bc-9901-09a4ab6da61c","Type":"ContainerStarted","Data":"16e98e0e4b5870dab35c4126c1d5f6b3f34689d1c137e5f66b0acb5f72c29760"} Feb 14 11:29:51 crc kubenswrapper[4904]: I0214 11:29:51.700710 4904 generic.go:334] "Generic (PLEG): container finished" podID="37dfca9c-dcd2-4620-b6c6-55225fc0693a" containerID="8bbdd4200ba9f703d1be47f0df0a36df1548e458965b28c4ef87d2a9757f4ec7" exitCode=0 Feb 14 11:29:51 crc kubenswrapper[4904]: I0214 11:29:51.700736 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" event={"ID":"37dfca9c-dcd2-4620-b6c6-55225fc0693a","Type":"ContainerDied","Data":"8bbdd4200ba9f703d1be47f0df0a36df1548e458965b28c4ef87d2a9757f4ec7"} Feb 14 11:29:51 crc kubenswrapper[4904]: I0214 11:29:51.874767 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.006274 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-config\") pod \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.006591 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-dns-svc\") pod \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.006648 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vltq\" (UniqueName: \"kubernetes.io/projected/37dfca9c-dcd2-4620-b6c6-55225fc0693a-kube-api-access-8vltq\") pod \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.006680 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-ovsdbserver-sb\") pod \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.006778 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-dns-swift-storage-0\") pod \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.006825 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-ovsdbserver-nb\") pod \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\" (UID: \"37dfca9c-dcd2-4620-b6c6-55225fc0693a\") " Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.012335 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37dfca9c-dcd2-4620-b6c6-55225fc0693a-kube-api-access-8vltq" (OuterVolumeSpecName: "kube-api-access-8vltq") pod "37dfca9c-dcd2-4620-b6c6-55225fc0693a" (UID: "37dfca9c-dcd2-4620-b6c6-55225fc0693a"). InnerVolumeSpecName "kube-api-access-8vltq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.056215 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "37dfca9c-dcd2-4620-b6c6-55225fc0693a" (UID: "37dfca9c-dcd2-4620-b6c6-55225fc0693a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.064442 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "37dfca9c-dcd2-4620-b6c6-55225fc0693a" (UID: "37dfca9c-dcd2-4620-b6c6-55225fc0693a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.077500 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "37dfca9c-dcd2-4620-b6c6-55225fc0693a" (UID: "37dfca9c-dcd2-4620-b6c6-55225fc0693a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.078140 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-config" (OuterVolumeSpecName: "config") pod "37dfca9c-dcd2-4620-b6c6-55225fc0693a" (UID: "37dfca9c-dcd2-4620-b6c6-55225fc0693a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.089893 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "37dfca9c-dcd2-4620-b6c6-55225fc0693a" (UID: "37dfca9c-dcd2-4620-b6c6-55225fc0693a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.113188 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.113224 4904 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.113251 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vltq\" (UniqueName: \"kubernetes.io/projected/37dfca9c-dcd2-4620-b6c6-55225fc0693a-kube-api-access-8vltq\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.113265 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.113274 4904 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.113282 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37dfca9c-dcd2-4620-b6c6-55225fc0693a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.710791 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4kbtk" event={"ID":"29681a87-9c4c-47fd-9834-c3cf2ebd79d3","Type":"ContainerStarted","Data":"2b034addfe9ba1b39e700ac2917b72a677f2e05b1ddc127dd0c3ed3a7f26a1c8"} Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.712873 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cbe346d4-5e03-40bc-9901-09a4ab6da61c","Type":"ContainerStarted","Data":"9fd7389b8bfca9a69f1f975410d382d61fbc474d1dda6eb6db9b817563154d21"} Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.714581 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" event={"ID":"37dfca9c-dcd2-4620-b6c6-55225fc0693a","Type":"ContainerDied","Data":"6db0443bff5f272d2c25dcaa444ad3f6f8350254aaca89108fb194ddb152e1c1"} Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.714621 4904 scope.go:117] "RemoveContainer" containerID="8bbdd4200ba9f703d1be47f0df0a36df1548e458965b28c4ef87d2a9757f4ec7" Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.714668 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.744049 4904 scope.go:117] "RemoveContainer" containerID="87c70ae353a105453d588f5c729f8c36d0bd7c3717964833db4e1d7e12a5b8fe" Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.746290 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-4kbtk" podStartSLOduration=3.746272256 podStartE2EDuration="3.746272256s" podCreationTimestamp="2026-02-14 11:29:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:29:52.738219034 +0000 UTC m=+1183.550983695" watchObservedRunningTime="2026-02-14 11:29:52.746272256 +0000 UTC m=+1183.559036907" Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.788720 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-9qq9d"] Feb 14 11:29:52 crc kubenswrapper[4904]: I0214 11:29:52.802251 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-9qq9d"] Feb 14 11:29:53 crc kubenswrapper[4904]: I0214 11:29:53.724376 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cbe346d4-5e03-40bc-9901-09a4ab6da61c","Type":"ContainerStarted","Data":"adb94f26579094b68ee541d0cf86dac683dc4f4245138e07c8b0a4d71b298972"} Feb 14 11:29:53 crc kubenswrapper[4904]: I0214 11:29:53.724676 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 14 11:29:53 crc kubenswrapper[4904]: I0214 11:29:53.745483 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.4619619209999999 podStartE2EDuration="4.745464533s" podCreationTimestamp="2026-02-14 11:29:49 +0000 UTC" firstStartedPulling="2026-02-14 11:29:49.869251887 +0000 UTC m=+1180.682016548" lastFinishedPulling="2026-02-14 11:29:53.152754499 +0000 UTC m=+1183.965519160" observedRunningTime="2026-02-14 11:29:53.744553698 +0000 UTC m=+1184.557318359" watchObservedRunningTime="2026-02-14 11:29:53.745464533 +0000 UTC m=+1184.558229194" Feb 14 11:29:53 crc kubenswrapper[4904]: I0214 11:29:53.859773 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37dfca9c-dcd2-4620-b6c6-55225fc0693a" path="/var/lib/kubelet/pods/37dfca9c-dcd2-4620-b6c6-55225fc0693a/volumes" Feb 14 11:29:56 crc kubenswrapper[4904]: I0214 11:29:56.537182 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-757b4f8459-9qq9d" podUID="37dfca9c-dcd2-4620-b6c6-55225fc0693a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.193:5353: i/o timeout" Feb 14 11:29:56 crc kubenswrapper[4904]: I0214 11:29:56.765012 4904 generic.go:334] "Generic (PLEG): container finished" podID="29681a87-9c4c-47fd-9834-c3cf2ebd79d3" containerID="2b034addfe9ba1b39e700ac2917b72a677f2e05b1ddc127dd0c3ed3a7f26a1c8" exitCode=0 Feb 14 11:29:56 crc kubenswrapper[4904]: I0214 11:29:56.765066 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4kbtk" event={"ID":"29681a87-9c4c-47fd-9834-c3cf2ebd79d3","Type":"ContainerDied","Data":"2b034addfe9ba1b39e700ac2917b72a677f2e05b1ddc127dd0c3ed3a7f26a1c8"} Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.070501 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.070861 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.159912 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4kbtk" Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.223517 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jvng\" (UniqueName: \"kubernetes.io/projected/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-kube-api-access-8jvng\") pod \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\" (UID: \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\") " Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.223731 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-combined-ca-bundle\") pod \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\" (UID: \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\") " Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.223766 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-config-data\") pod \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\" (UID: \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\") " Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.223847 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-scripts\") pod \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\" (UID: \"29681a87-9c4c-47fd-9834-c3cf2ebd79d3\") " Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.244049 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-scripts" (OuterVolumeSpecName: "scripts") pod "29681a87-9c4c-47fd-9834-c3cf2ebd79d3" (UID: "29681a87-9c4c-47fd-9834-c3cf2ebd79d3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.244146 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-kube-api-access-8jvng" (OuterVolumeSpecName: "kube-api-access-8jvng") pod "29681a87-9c4c-47fd-9834-c3cf2ebd79d3" (UID: "29681a87-9c4c-47fd-9834-c3cf2ebd79d3"). InnerVolumeSpecName "kube-api-access-8jvng". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.256610 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "29681a87-9c4c-47fd-9834-c3cf2ebd79d3" (UID: "29681a87-9c4c-47fd-9834-c3cf2ebd79d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.263758 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-config-data" (OuterVolumeSpecName: "config-data") pod "29681a87-9c4c-47fd-9834-c3cf2ebd79d3" (UID: "29681a87-9c4c-47fd-9834-c3cf2ebd79d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.327022 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.327052 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.327062 4904 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-scripts\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.327071 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jvng\" (UniqueName: \"kubernetes.io/projected/29681a87-9c4c-47fd-9834-c3cf2ebd79d3-kube-api-access-8jvng\") on node \"crc\" DevicePath \"\"" Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.783123 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4kbtk" event={"ID":"29681a87-9c4c-47fd-9834-c3cf2ebd79d3","Type":"ContainerDied","Data":"ae80c5c61032338d6856d97291ad28aac95f0950ebecae35e2308b3d9c164820"} Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.783164 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae80c5c61032338d6856d97291ad28aac95f0950ebecae35e2308b3d9c164820" Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.783222 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4kbtk" Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.977336 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.977623 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="de9dd864-0953-44cf-ad93-5284648538f2" containerName="nova-api-api" containerID="cri-o://7709ecf24b0951207660c1134dbe1ed2247dab2b32054996f1d80f996e404312" gracePeriod=30 Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.977618 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="de9dd864-0953-44cf-ad93-5284648538f2" containerName="nova-api-log" containerID="cri-o://000e175ba2a360712075d7b0cd3d6fc23f990e1253cbfb4384baefc541503401" gracePeriod=30 Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.987157 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="de9dd864-0953-44cf-ad93-5284648538f2" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": EOF" Feb 14 11:29:58 crc kubenswrapper[4904]: I0214 11:29:58.987326 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="de9dd864-0953-44cf-ad93-5284648538f2" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": EOF" Feb 14 11:29:59 crc kubenswrapper[4904]: I0214 11:29:59.007278 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 14 11:29:59 crc kubenswrapper[4904]: I0214 11:29:59.008172 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="395d1b56-d8ac-4d0c-9b35-17950f8e1da3" containerName="nova-scheduler-scheduler" containerID="cri-o://4bb40494658f13cf27aa4e454f8af6af19851d098b00fa02d313310db4794fb9" gracePeriod=30 Feb 14 11:29:59 crc kubenswrapper[4904]: I0214 11:29:59.026587 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:29:59 crc kubenswrapper[4904]: I0214 11:29:59.027480 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5b248689-1264-4fac-a77e-60f6494317f3" containerName="nova-metadata-log" containerID="cri-o://68b584bdd5beda2ef787496cdb9858bdb40b347ec9054a909d3342040e3d2234" gracePeriod=30 Feb 14 11:29:59 crc kubenswrapper[4904]: I0214 11:29:59.027597 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5b248689-1264-4fac-a77e-60f6494317f3" containerName="nova-metadata-metadata" containerID="cri-o://929b4ee3de69c4956fb06ed9107d192c5bc4eada12f4ccf24ecff2018abc128f" gracePeriod=30 Feb 14 11:29:59 crc kubenswrapper[4904]: I0214 11:29:59.793162 4904 generic.go:334] "Generic (PLEG): container finished" podID="de9dd864-0953-44cf-ad93-5284648538f2" containerID="000e175ba2a360712075d7b0cd3d6fc23f990e1253cbfb4384baefc541503401" exitCode=143 Feb 14 11:29:59 crc kubenswrapper[4904]: I0214 11:29:59.793198 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de9dd864-0953-44cf-ad93-5284648538f2","Type":"ContainerDied","Data":"000e175ba2a360712075d7b0cd3d6fc23f990e1253cbfb4384baefc541503401"} Feb 14 11:29:59 crc kubenswrapper[4904]: I0214 11:29:59.795061 4904 generic.go:334] "Generic (PLEG): container finished" podID="5b248689-1264-4fac-a77e-60f6494317f3" containerID="68b584bdd5beda2ef787496cdb9858bdb40b347ec9054a909d3342040e3d2234" exitCode=143 Feb 14 11:29:59 crc kubenswrapper[4904]: I0214 11:29:59.795106 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b248689-1264-4fac-a77e-60f6494317f3","Type":"ContainerDied","Data":"68b584bdd5beda2ef787496cdb9858bdb40b347ec9054a909d3342040e3d2234"} Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.159018 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl"] Feb 14 11:30:00 crc kubenswrapper[4904]: E0214 11:30:00.159693 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37dfca9c-dcd2-4620-b6c6-55225fc0693a" containerName="dnsmasq-dns" Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.159710 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="37dfca9c-dcd2-4620-b6c6-55225fc0693a" containerName="dnsmasq-dns" Feb 14 11:30:00 crc kubenswrapper[4904]: E0214 11:30:00.159723 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37dfca9c-dcd2-4620-b6c6-55225fc0693a" containerName="init" Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.159729 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="37dfca9c-dcd2-4620-b6c6-55225fc0693a" containerName="init" Feb 14 11:30:00 crc kubenswrapper[4904]: E0214 11:30:00.159742 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29681a87-9c4c-47fd-9834-c3cf2ebd79d3" containerName="nova-manage" Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.159752 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="29681a87-9c4c-47fd-9834-c3cf2ebd79d3" containerName="nova-manage" Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.159969 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="37dfca9c-dcd2-4620-b6c6-55225fc0693a" containerName="dnsmasq-dns" Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.159997 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="29681a87-9c4c-47fd-9834-c3cf2ebd79d3" containerName="nova-manage" Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.160911 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl" Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.167291 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.168132 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.169639 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl"] Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.270454 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw25j\" (UniqueName: \"kubernetes.io/projected/71c31cb0-4057-4263-96d8-8938381f2b5b-kube-api-access-lw25j\") pod \"collect-profiles-29517810-4d9dl\" (UID: \"71c31cb0-4057-4263-96d8-8938381f2b5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl" Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.270517 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71c31cb0-4057-4263-96d8-8938381f2b5b-config-volume\") pod \"collect-profiles-29517810-4d9dl\" (UID: \"71c31cb0-4057-4263-96d8-8938381f2b5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl" Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.270613 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71c31cb0-4057-4263-96d8-8938381f2b5b-secret-volume\") pod \"collect-profiles-29517810-4d9dl\" (UID: \"71c31cb0-4057-4263-96d8-8938381f2b5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl" Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.372446 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw25j\" (UniqueName: \"kubernetes.io/projected/71c31cb0-4057-4263-96d8-8938381f2b5b-kube-api-access-lw25j\") pod \"collect-profiles-29517810-4d9dl\" (UID: \"71c31cb0-4057-4263-96d8-8938381f2b5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl" Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.372513 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71c31cb0-4057-4263-96d8-8938381f2b5b-config-volume\") pod \"collect-profiles-29517810-4d9dl\" (UID: \"71c31cb0-4057-4263-96d8-8938381f2b5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl" Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.372584 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71c31cb0-4057-4263-96d8-8938381f2b5b-secret-volume\") pod \"collect-profiles-29517810-4d9dl\" (UID: \"71c31cb0-4057-4263-96d8-8938381f2b5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl" Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.373637 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71c31cb0-4057-4263-96d8-8938381f2b5b-config-volume\") pod \"collect-profiles-29517810-4d9dl\" (UID: \"71c31cb0-4057-4263-96d8-8938381f2b5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl" Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.381745 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71c31cb0-4057-4263-96d8-8938381f2b5b-secret-volume\") pod \"collect-profiles-29517810-4d9dl\" (UID: \"71c31cb0-4057-4263-96d8-8938381f2b5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl" Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.398760 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw25j\" (UniqueName: \"kubernetes.io/projected/71c31cb0-4057-4263-96d8-8938381f2b5b-kube-api-access-lw25j\") pod \"collect-profiles-29517810-4d9dl\" (UID: \"71c31cb0-4057-4263-96d8-8938381f2b5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl" Feb 14 11:30:00 crc kubenswrapper[4904]: I0214 11:30:00.480394 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl" Feb 14 11:30:00 crc kubenswrapper[4904]: E0214 11:30:00.656576 4904 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4bb40494658f13cf27aa4e454f8af6af19851d098b00fa02d313310db4794fb9" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 14 11:30:00 crc kubenswrapper[4904]: E0214 11:30:00.660624 4904 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4bb40494658f13cf27aa4e454f8af6af19851d098b00fa02d313310db4794fb9" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 14 11:30:00 crc kubenswrapper[4904]: E0214 11:30:00.668158 4904 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4bb40494658f13cf27aa4e454f8af6af19851d098b00fa02d313310db4794fb9" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 14 11:30:00 crc kubenswrapper[4904]: E0214 11:30:00.668218 4904 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="395d1b56-d8ac-4d0c-9b35-17950f8e1da3" containerName="nova-scheduler-scheduler" Feb 14 11:30:01 crc kubenswrapper[4904]: I0214 11:30:00.999827 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl"] Feb 14 11:30:01 crc kubenswrapper[4904]: I0214 11:30:01.810528 4904 generic.go:334] "Generic (PLEG): container finished" podID="71c31cb0-4057-4263-96d8-8938381f2b5b" containerID="b0ce3b7f36a57048b805388454f435ef2c7144169b88c9384c3de7802534c066" exitCode=0 Feb 14 11:30:01 crc kubenswrapper[4904]: I0214 11:30:01.810634 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl" event={"ID":"71c31cb0-4057-4263-96d8-8938381f2b5b","Type":"ContainerDied","Data":"b0ce3b7f36a57048b805388454f435ef2c7144169b88c9384c3de7802534c066"} Feb 14 11:30:01 crc kubenswrapper[4904]: I0214 11:30:01.810862 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl" event={"ID":"71c31cb0-4057-4263-96d8-8938381f2b5b","Type":"ContainerStarted","Data":"62af5613645173c25c5a9450cdda5e94f05a56bc082a9c902dcc7da7f2342f10"} Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.343751 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="5b248689-1264-4fac-a77e-60f6494317f3" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": read tcp 10.217.0.2:45424->10.217.0.198:8775: read: connection reset by peer" Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.344164 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="5b248689-1264-4fac-a77e-60f6494317f3" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.198:8775/\": read tcp 10.217.0.2:45438->10.217.0.198:8775: read: connection reset by peer" Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.800701 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.811579 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b248689-1264-4fac-a77e-60f6494317f3-combined-ca-bundle\") pod \"5b248689-1264-4fac-a77e-60f6494317f3\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.811670 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5t7gf\" (UniqueName: \"kubernetes.io/projected/5b248689-1264-4fac-a77e-60f6494317f3-kube-api-access-5t7gf\") pod \"5b248689-1264-4fac-a77e-60f6494317f3\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.811719 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b248689-1264-4fac-a77e-60f6494317f3-config-data\") pod \"5b248689-1264-4fac-a77e-60f6494317f3\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.811757 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b248689-1264-4fac-a77e-60f6494317f3-nova-metadata-tls-certs\") pod \"5b248689-1264-4fac-a77e-60f6494317f3\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.811938 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b248689-1264-4fac-a77e-60f6494317f3-logs\") pod \"5b248689-1264-4fac-a77e-60f6494317f3\" (UID: \"5b248689-1264-4fac-a77e-60f6494317f3\") " Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.812364 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b248689-1264-4fac-a77e-60f6494317f3-logs" (OuterVolumeSpecName: "logs") pod "5b248689-1264-4fac-a77e-60f6494317f3" (UID: "5b248689-1264-4fac-a77e-60f6494317f3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.818446 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b248689-1264-4fac-a77e-60f6494317f3-kube-api-access-5t7gf" (OuterVolumeSpecName: "kube-api-access-5t7gf") pod "5b248689-1264-4fac-a77e-60f6494317f3" (UID: "5b248689-1264-4fac-a77e-60f6494317f3"). InnerVolumeSpecName "kube-api-access-5t7gf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.825602 4904 generic.go:334] "Generic (PLEG): container finished" podID="5b248689-1264-4fac-a77e-60f6494317f3" containerID="929b4ee3de69c4956fb06ed9107d192c5bc4eada12f4ccf24ecff2018abc128f" exitCode=0 Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.826007 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.826704 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b248689-1264-4fac-a77e-60f6494317f3","Type":"ContainerDied","Data":"929b4ee3de69c4956fb06ed9107d192c5bc4eada12f4ccf24ecff2018abc128f"} Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.828466 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b248689-1264-4fac-a77e-60f6494317f3","Type":"ContainerDied","Data":"1c7bfe4372346aceb055b05185212db7e067b8ed80075647c60fb22c06f124bc"} Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.828535 4904 scope.go:117] "RemoveContainer" containerID="929b4ee3de69c4956fb06ed9107d192c5bc4eada12f4ccf24ecff2018abc128f" Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.868947 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b248689-1264-4fac-a77e-60f6494317f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b248689-1264-4fac-a77e-60f6494317f3" (UID: "5b248689-1264-4fac-a77e-60f6494317f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.898212 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b248689-1264-4fac-a77e-60f6494317f3-config-data" (OuterVolumeSpecName: "config-data") pod "5b248689-1264-4fac-a77e-60f6494317f3" (UID: "5b248689-1264-4fac-a77e-60f6494317f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.914576 4904 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b248689-1264-4fac-a77e-60f6494317f3-logs\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.914623 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b248689-1264-4fac-a77e-60f6494317f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.914634 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5t7gf\" (UniqueName: \"kubernetes.io/projected/5b248689-1264-4fac-a77e-60f6494317f3-kube-api-access-5t7gf\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.914642 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b248689-1264-4fac-a77e-60f6494317f3-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.937983 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b248689-1264-4fac-a77e-60f6494317f3-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "5b248689-1264-4fac-a77e-60f6494317f3" (UID: "5b248689-1264-4fac-a77e-60f6494317f3"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.950845 4904 scope.go:117] "RemoveContainer" containerID="68b584bdd5beda2ef787496cdb9858bdb40b347ec9054a909d3342040e3d2234" Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.975853 4904 scope.go:117] "RemoveContainer" containerID="929b4ee3de69c4956fb06ed9107d192c5bc4eada12f4ccf24ecff2018abc128f" Feb 14 11:30:02 crc kubenswrapper[4904]: E0214 11:30:02.976419 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"929b4ee3de69c4956fb06ed9107d192c5bc4eada12f4ccf24ecff2018abc128f\": container with ID starting with 929b4ee3de69c4956fb06ed9107d192c5bc4eada12f4ccf24ecff2018abc128f not found: ID does not exist" containerID="929b4ee3de69c4956fb06ed9107d192c5bc4eada12f4ccf24ecff2018abc128f" Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.976445 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"929b4ee3de69c4956fb06ed9107d192c5bc4eada12f4ccf24ecff2018abc128f"} err="failed to get container status \"929b4ee3de69c4956fb06ed9107d192c5bc4eada12f4ccf24ecff2018abc128f\": rpc error: code = NotFound desc = could not find container \"929b4ee3de69c4956fb06ed9107d192c5bc4eada12f4ccf24ecff2018abc128f\": container with ID starting with 929b4ee3de69c4956fb06ed9107d192c5bc4eada12f4ccf24ecff2018abc128f not found: ID does not exist" Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.976463 4904 scope.go:117] "RemoveContainer" containerID="68b584bdd5beda2ef787496cdb9858bdb40b347ec9054a909d3342040e3d2234" Feb 14 11:30:02 crc kubenswrapper[4904]: E0214 11:30:02.976866 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68b584bdd5beda2ef787496cdb9858bdb40b347ec9054a909d3342040e3d2234\": container with ID starting with 68b584bdd5beda2ef787496cdb9858bdb40b347ec9054a909d3342040e3d2234 not found: ID does not exist" containerID="68b584bdd5beda2ef787496cdb9858bdb40b347ec9054a909d3342040e3d2234" Feb 14 11:30:02 crc kubenswrapper[4904]: I0214 11:30:02.976890 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68b584bdd5beda2ef787496cdb9858bdb40b347ec9054a909d3342040e3d2234"} err="failed to get container status \"68b584bdd5beda2ef787496cdb9858bdb40b347ec9054a909d3342040e3d2234\": rpc error: code = NotFound desc = could not find container \"68b584bdd5beda2ef787496cdb9858bdb40b347ec9054a909d3342040e3d2234\": container with ID starting with 68b584bdd5beda2ef787496cdb9858bdb40b347ec9054a909d3342040e3d2234 not found: ID does not exist" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.016183 4904 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b248689-1264-4fac-a77e-60f6494317f3-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.161404 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.171741 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.185258 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.187110 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:30:03 crc kubenswrapper[4904]: E0214 11:30:03.187574 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71c31cb0-4057-4263-96d8-8938381f2b5b" containerName="collect-profiles" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.187598 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="71c31cb0-4057-4263-96d8-8938381f2b5b" containerName="collect-profiles" Feb 14 11:30:03 crc kubenswrapper[4904]: E0214 11:30:03.187631 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b248689-1264-4fac-a77e-60f6494317f3" containerName="nova-metadata-metadata" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.187637 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b248689-1264-4fac-a77e-60f6494317f3" containerName="nova-metadata-metadata" Feb 14 11:30:03 crc kubenswrapper[4904]: E0214 11:30:03.187653 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b248689-1264-4fac-a77e-60f6494317f3" containerName="nova-metadata-log" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.187659 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b248689-1264-4fac-a77e-60f6494317f3" containerName="nova-metadata-log" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.190904 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="71c31cb0-4057-4263-96d8-8938381f2b5b" containerName="collect-profiles" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.190952 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b248689-1264-4fac-a77e-60f6494317f3" containerName="nova-metadata-log" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.190978 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b248689-1264-4fac-a77e-60f6494317f3" containerName="nova-metadata-metadata" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.192106 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.195930 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.196216 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.210316 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.322325 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lw25j\" (UniqueName: \"kubernetes.io/projected/71c31cb0-4057-4263-96d8-8938381f2b5b-kube-api-access-lw25j\") pod \"71c31cb0-4057-4263-96d8-8938381f2b5b\" (UID: \"71c31cb0-4057-4263-96d8-8938381f2b5b\") " Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.322398 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71c31cb0-4057-4263-96d8-8938381f2b5b-secret-volume\") pod \"71c31cb0-4057-4263-96d8-8938381f2b5b\" (UID: \"71c31cb0-4057-4263-96d8-8938381f2b5b\") " Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.322540 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71c31cb0-4057-4263-96d8-8938381f2b5b-config-volume\") pod \"71c31cb0-4057-4263-96d8-8938381f2b5b\" (UID: \"71c31cb0-4057-4263-96d8-8938381f2b5b\") " Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.322902 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/aba968e8-e711-4cbc-aea2-6e68c7c9a27c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"aba968e8-e711-4cbc-aea2-6e68c7c9a27c\") " pod="openstack/nova-metadata-0" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.322960 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aba968e8-e711-4cbc-aea2-6e68c7c9a27c-config-data\") pod \"nova-metadata-0\" (UID: \"aba968e8-e711-4cbc-aea2-6e68c7c9a27c\") " pod="openstack/nova-metadata-0" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.323183 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aba968e8-e711-4cbc-aea2-6e68c7c9a27c-logs\") pod \"nova-metadata-0\" (UID: \"aba968e8-e711-4cbc-aea2-6e68c7c9a27c\") " pod="openstack/nova-metadata-0" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.323380 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aba968e8-e711-4cbc-aea2-6e68c7c9a27c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"aba968e8-e711-4cbc-aea2-6e68c7c9a27c\") " pod="openstack/nova-metadata-0" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.323424 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71c31cb0-4057-4263-96d8-8938381f2b5b-config-volume" (OuterVolumeSpecName: "config-volume") pod "71c31cb0-4057-4263-96d8-8938381f2b5b" (UID: "71c31cb0-4057-4263-96d8-8938381f2b5b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.323456 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nlbl\" (UniqueName: \"kubernetes.io/projected/aba968e8-e711-4cbc-aea2-6e68c7c9a27c-kube-api-access-2nlbl\") pod \"nova-metadata-0\" (UID: \"aba968e8-e711-4cbc-aea2-6e68c7c9a27c\") " pod="openstack/nova-metadata-0" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.323588 4904 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/71c31cb0-4057-4263-96d8-8938381f2b5b-config-volume\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.325575 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71c31cb0-4057-4263-96d8-8938381f2b5b-kube-api-access-lw25j" (OuterVolumeSpecName: "kube-api-access-lw25j") pod "71c31cb0-4057-4263-96d8-8938381f2b5b" (UID: "71c31cb0-4057-4263-96d8-8938381f2b5b"). InnerVolumeSpecName "kube-api-access-lw25j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.326695 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71c31cb0-4057-4263-96d8-8938381f2b5b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "71c31cb0-4057-4263-96d8-8938381f2b5b" (UID: "71c31cb0-4057-4263-96d8-8938381f2b5b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.425274 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aba968e8-e711-4cbc-aea2-6e68c7c9a27c-logs\") pod \"nova-metadata-0\" (UID: \"aba968e8-e711-4cbc-aea2-6e68c7c9a27c\") " pod="openstack/nova-metadata-0" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.425629 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aba968e8-e711-4cbc-aea2-6e68c7c9a27c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"aba968e8-e711-4cbc-aea2-6e68c7c9a27c\") " pod="openstack/nova-metadata-0" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.425664 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nlbl\" (UniqueName: \"kubernetes.io/projected/aba968e8-e711-4cbc-aea2-6e68c7c9a27c-kube-api-access-2nlbl\") pod \"nova-metadata-0\" (UID: \"aba968e8-e711-4cbc-aea2-6e68c7c9a27c\") " pod="openstack/nova-metadata-0" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.425712 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/aba968e8-e711-4cbc-aea2-6e68c7c9a27c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"aba968e8-e711-4cbc-aea2-6e68c7c9a27c\") " pod="openstack/nova-metadata-0" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.425760 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aba968e8-e711-4cbc-aea2-6e68c7c9a27c-config-data\") pod \"nova-metadata-0\" (UID: \"aba968e8-e711-4cbc-aea2-6e68c7c9a27c\") " pod="openstack/nova-metadata-0" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.425820 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lw25j\" (UniqueName: \"kubernetes.io/projected/71c31cb0-4057-4263-96d8-8938381f2b5b-kube-api-access-lw25j\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.425851 4904 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/71c31cb0-4057-4263-96d8-8938381f2b5b-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.426743 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aba968e8-e711-4cbc-aea2-6e68c7c9a27c-logs\") pod \"nova-metadata-0\" (UID: \"aba968e8-e711-4cbc-aea2-6e68c7c9a27c\") " pod="openstack/nova-metadata-0" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.428916 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aba968e8-e711-4cbc-aea2-6e68c7c9a27c-config-data\") pod \"nova-metadata-0\" (UID: \"aba968e8-e711-4cbc-aea2-6e68c7c9a27c\") " pod="openstack/nova-metadata-0" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.429083 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aba968e8-e711-4cbc-aea2-6e68c7c9a27c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"aba968e8-e711-4cbc-aea2-6e68c7c9a27c\") " pod="openstack/nova-metadata-0" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.429248 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/aba968e8-e711-4cbc-aea2-6e68c7c9a27c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"aba968e8-e711-4cbc-aea2-6e68c7c9a27c\") " pod="openstack/nova-metadata-0" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.442592 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nlbl\" (UniqueName: \"kubernetes.io/projected/aba968e8-e711-4cbc-aea2-6e68c7c9a27c-kube-api-access-2nlbl\") pod \"nova-metadata-0\" (UID: \"aba968e8-e711-4cbc-aea2-6e68c7c9a27c\") " pod="openstack/nova-metadata-0" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.542564 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.837759 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.853014 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b248689-1264-4fac-a77e-60f6494317f3" path="/var/lib/kubelet/pods/5b248689-1264-4fac-a77e-60f6494317f3/volumes" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.853870 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl" event={"ID":"71c31cb0-4057-4263-96d8-8938381f2b5b","Type":"ContainerDied","Data":"62af5613645173c25c5a9450cdda5e94f05a56bc082a9c902dcc7da7f2342f10"} Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.853976 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62af5613645173c25c5a9450cdda5e94f05a56bc082a9c902dcc7da7f2342f10" Feb 14 11:30:03 crc kubenswrapper[4904]: I0214 11:30:03.996300 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.403227 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.548625 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/395d1b56-d8ac-4d0c-9b35-17950f8e1da3-combined-ca-bundle\") pod \"395d1b56-d8ac-4d0c-9b35-17950f8e1da3\" (UID: \"395d1b56-d8ac-4d0c-9b35-17950f8e1da3\") " Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.548829 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/395d1b56-d8ac-4d0c-9b35-17950f8e1da3-config-data\") pod \"395d1b56-d8ac-4d0c-9b35-17950f8e1da3\" (UID: \"395d1b56-d8ac-4d0c-9b35-17950f8e1da3\") " Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.550597 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzpjd\" (UniqueName: \"kubernetes.io/projected/395d1b56-d8ac-4d0c-9b35-17950f8e1da3-kube-api-access-pzpjd\") pod \"395d1b56-d8ac-4d0c-9b35-17950f8e1da3\" (UID: \"395d1b56-d8ac-4d0c-9b35-17950f8e1da3\") " Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.559793 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/395d1b56-d8ac-4d0c-9b35-17950f8e1da3-kube-api-access-pzpjd" (OuterVolumeSpecName: "kube-api-access-pzpjd") pod "395d1b56-d8ac-4d0c-9b35-17950f8e1da3" (UID: "395d1b56-d8ac-4d0c-9b35-17950f8e1da3"). InnerVolumeSpecName "kube-api-access-pzpjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.576068 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/395d1b56-d8ac-4d0c-9b35-17950f8e1da3-config-data" (OuterVolumeSpecName: "config-data") pod "395d1b56-d8ac-4d0c-9b35-17950f8e1da3" (UID: "395d1b56-d8ac-4d0c-9b35-17950f8e1da3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.599151 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/395d1b56-d8ac-4d0c-9b35-17950f8e1da3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "395d1b56-d8ac-4d0c-9b35-17950f8e1da3" (UID: "395d1b56-d8ac-4d0c-9b35-17950f8e1da3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.652143 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/395d1b56-d8ac-4d0c-9b35-17950f8e1da3-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.652341 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzpjd\" (UniqueName: \"kubernetes.io/projected/395d1b56-d8ac-4d0c-9b35-17950f8e1da3-kube-api-access-pzpjd\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.652400 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/395d1b56-d8ac-4d0c-9b35-17950f8e1da3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.857553 4904 generic.go:334] "Generic (PLEG): container finished" podID="395d1b56-d8ac-4d0c-9b35-17950f8e1da3" containerID="4bb40494658f13cf27aa4e454f8af6af19851d098b00fa02d313310db4794fb9" exitCode=0 Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.857615 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"395d1b56-d8ac-4d0c-9b35-17950f8e1da3","Type":"ContainerDied","Data":"4bb40494658f13cf27aa4e454f8af6af19851d098b00fa02d313310db4794fb9"} Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.857597 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.857645 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"395d1b56-d8ac-4d0c-9b35-17950f8e1da3","Type":"ContainerDied","Data":"1c8fcef3391f1b88c02d3a7be07e4d8e79d14814f406e84cc66b77590bf81bdb"} Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.857692 4904 scope.go:117] "RemoveContainer" containerID="4bb40494658f13cf27aa4e454f8af6af19851d098b00fa02d313310db4794fb9" Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.860788 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aba968e8-e711-4cbc-aea2-6e68c7c9a27c","Type":"ContainerStarted","Data":"a6a3875def388ff2625070ce7eebc654eb36974cddbd56e6b7f950a12cd83d12"} Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.860818 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aba968e8-e711-4cbc-aea2-6e68c7c9a27c","Type":"ContainerStarted","Data":"70e3950916ac8b88d72ba3e29c9450e51f5c1408da0dff55d9033fb21f1eb036"} Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.860848 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aba968e8-e711-4cbc-aea2-6e68c7c9a27c","Type":"ContainerStarted","Data":"eacd3f57fb67297c33cb01549757edb5f29336ed4c715d4cef8493ca149ced6e"} Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.894625 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.8946033770000001 podStartE2EDuration="1.894603377s" podCreationTimestamp="2026-02-14 11:30:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:30:04.887107011 +0000 UTC m=+1195.699871672" watchObservedRunningTime="2026-02-14 11:30:04.894603377 +0000 UTC m=+1195.707368038" Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.913865 4904 scope.go:117] "RemoveContainer" containerID="4bb40494658f13cf27aa4e454f8af6af19851d098b00fa02d313310db4794fb9" Feb 14 11:30:04 crc kubenswrapper[4904]: E0214 11:30:04.914394 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bb40494658f13cf27aa4e454f8af6af19851d098b00fa02d313310db4794fb9\": container with ID starting with 4bb40494658f13cf27aa4e454f8af6af19851d098b00fa02d313310db4794fb9 not found: ID does not exist" containerID="4bb40494658f13cf27aa4e454f8af6af19851d098b00fa02d313310db4794fb9" Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.914438 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bb40494658f13cf27aa4e454f8af6af19851d098b00fa02d313310db4794fb9"} err="failed to get container status \"4bb40494658f13cf27aa4e454f8af6af19851d098b00fa02d313310db4794fb9\": rpc error: code = NotFound desc = could not find container \"4bb40494658f13cf27aa4e454f8af6af19851d098b00fa02d313310db4794fb9\": container with ID starting with 4bb40494658f13cf27aa4e454f8af6af19851d098b00fa02d313310db4794fb9 not found: ID does not exist" Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.922205 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.937461 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.947653 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 14 11:30:04 crc kubenswrapper[4904]: E0214 11:30:04.948070 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="395d1b56-d8ac-4d0c-9b35-17950f8e1da3" containerName="nova-scheduler-scheduler" Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.948091 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="395d1b56-d8ac-4d0c-9b35-17950f8e1da3" containerName="nova-scheduler-scheduler" Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.948262 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="395d1b56-d8ac-4d0c-9b35-17950f8e1da3" containerName="nova-scheduler-scheduler" Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.948822 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.950574 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 14 11:30:04 crc kubenswrapper[4904]: I0214 11:30:04.956540 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.059704 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q84m6\" (UniqueName: \"kubernetes.io/projected/dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5-kube-api-access-q84m6\") pod \"nova-scheduler-0\" (UID: \"dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5\") " pod="openstack/nova-scheduler-0" Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.060155 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5-config-data\") pod \"nova-scheduler-0\" (UID: \"dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5\") " pod="openstack/nova-scheduler-0" Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.060187 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5\") " pod="openstack/nova-scheduler-0" Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.163653 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5-config-data\") pod \"nova-scheduler-0\" (UID: \"dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5\") " pod="openstack/nova-scheduler-0" Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.163692 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5\") " pod="openstack/nova-scheduler-0" Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.163796 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q84m6\" (UniqueName: \"kubernetes.io/projected/dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5-kube-api-access-q84m6\") pod \"nova-scheduler-0\" (UID: \"dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5\") " pod="openstack/nova-scheduler-0" Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.169322 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5\") " pod="openstack/nova-scheduler-0" Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.169495 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5-config-data\") pod \"nova-scheduler-0\" (UID: \"dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5\") " pod="openstack/nova-scheduler-0" Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.181988 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q84m6\" (UniqueName: \"kubernetes.io/projected/dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5-kube-api-access-q84m6\") pod \"nova-scheduler-0\" (UID: \"dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5\") " pod="openstack/nova-scheduler-0" Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.267333 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.719973 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.849120 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="395d1b56-d8ac-4d0c-9b35-17950f8e1da3" path="/var/lib/kubelet/pods/395d1b56-d8ac-4d0c-9b35-17950f8e1da3/volumes" Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.874250 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5","Type":"ContainerStarted","Data":"b96c2b53ccfb9ddcb074f4123763c3ebd2927aa9b50a76eab11e26ae026892ea"} Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.877235 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.881405 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de9dd864-0953-44cf-ad93-5284648538f2","Type":"ContainerDied","Data":"7709ecf24b0951207660c1134dbe1ed2247dab2b32054996f1d80f996e404312"} Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.881382 4904 generic.go:334] "Generic (PLEG): container finished" podID="de9dd864-0953-44cf-ad93-5284648538f2" containerID="7709ecf24b0951207660c1134dbe1ed2247dab2b32054996f1d80f996e404312" exitCode=0 Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.881447 4904 scope.go:117] "RemoveContainer" containerID="7709ecf24b0951207660c1134dbe1ed2247dab2b32054996f1d80f996e404312" Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.881774 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de9dd864-0953-44cf-ad93-5284648538f2","Type":"ContainerDied","Data":"7d0b6b5f18867da24a2a2a9fe18f01918092ea279ea17231f2ea8e7891a5cb65"} Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.946745 4904 scope.go:117] "RemoveContainer" containerID="000e175ba2a360712075d7b0cd3d6fc23f990e1253cbfb4384baefc541503401" Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.966716 4904 scope.go:117] "RemoveContainer" containerID="7709ecf24b0951207660c1134dbe1ed2247dab2b32054996f1d80f996e404312" Feb 14 11:30:05 crc kubenswrapper[4904]: E0214 11:30:05.967623 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7709ecf24b0951207660c1134dbe1ed2247dab2b32054996f1d80f996e404312\": container with ID starting with 7709ecf24b0951207660c1134dbe1ed2247dab2b32054996f1d80f996e404312 not found: ID does not exist" containerID="7709ecf24b0951207660c1134dbe1ed2247dab2b32054996f1d80f996e404312" Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.967664 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7709ecf24b0951207660c1134dbe1ed2247dab2b32054996f1d80f996e404312"} err="failed to get container status \"7709ecf24b0951207660c1134dbe1ed2247dab2b32054996f1d80f996e404312\": rpc error: code = NotFound desc = could not find container \"7709ecf24b0951207660c1134dbe1ed2247dab2b32054996f1d80f996e404312\": container with ID starting with 7709ecf24b0951207660c1134dbe1ed2247dab2b32054996f1d80f996e404312 not found: ID does not exist" Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.967710 4904 scope.go:117] "RemoveContainer" containerID="000e175ba2a360712075d7b0cd3d6fc23f990e1253cbfb4384baefc541503401" Feb 14 11:30:05 crc kubenswrapper[4904]: E0214 11:30:05.968177 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"000e175ba2a360712075d7b0cd3d6fc23f990e1253cbfb4384baefc541503401\": container with ID starting with 000e175ba2a360712075d7b0cd3d6fc23f990e1253cbfb4384baefc541503401 not found: ID does not exist" containerID="000e175ba2a360712075d7b0cd3d6fc23f990e1253cbfb4384baefc541503401" Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.968201 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"000e175ba2a360712075d7b0cd3d6fc23f990e1253cbfb4384baefc541503401"} err="failed to get container status \"000e175ba2a360712075d7b0cd3d6fc23f990e1253cbfb4384baefc541503401\": rpc error: code = NotFound desc = could not find container \"000e175ba2a360712075d7b0cd3d6fc23f990e1253cbfb4384baefc541503401\": container with ID starting with 000e175ba2a360712075d7b0cd3d6fc23f990e1253cbfb4384baefc541503401 not found: ID does not exist" Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.981512 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khd78\" (UniqueName: \"kubernetes.io/projected/de9dd864-0953-44cf-ad93-5284648538f2-kube-api-access-khd78\") pod \"de9dd864-0953-44cf-ad93-5284648538f2\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.981604 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-config-data\") pod \"de9dd864-0953-44cf-ad93-5284648538f2\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.981656 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de9dd864-0953-44cf-ad93-5284648538f2-logs\") pod \"de9dd864-0953-44cf-ad93-5284648538f2\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.981720 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-internal-tls-certs\") pod \"de9dd864-0953-44cf-ad93-5284648538f2\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.981812 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-combined-ca-bundle\") pod \"de9dd864-0953-44cf-ad93-5284648538f2\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.981847 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-public-tls-certs\") pod \"de9dd864-0953-44cf-ad93-5284648538f2\" (UID: \"de9dd864-0953-44cf-ad93-5284648538f2\") " Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.983453 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de9dd864-0953-44cf-ad93-5284648538f2-logs" (OuterVolumeSpecName: "logs") pod "de9dd864-0953-44cf-ad93-5284648538f2" (UID: "de9dd864-0953-44cf-ad93-5284648538f2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:30:05 crc kubenswrapper[4904]: I0214 11:30:05.986237 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de9dd864-0953-44cf-ad93-5284648538f2-kube-api-access-khd78" (OuterVolumeSpecName: "kube-api-access-khd78") pod "de9dd864-0953-44cf-ad93-5284648538f2" (UID: "de9dd864-0953-44cf-ad93-5284648538f2"). InnerVolumeSpecName "kube-api-access-khd78". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.006949 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-config-data" (OuterVolumeSpecName: "config-data") pod "de9dd864-0953-44cf-ad93-5284648538f2" (UID: "de9dd864-0953-44cf-ad93-5284648538f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.008462 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de9dd864-0953-44cf-ad93-5284648538f2" (UID: "de9dd864-0953-44cf-ad93-5284648538f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.030069 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "de9dd864-0953-44cf-ad93-5284648538f2" (UID: "de9dd864-0953-44cf-ad93-5284648538f2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.031002 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "de9dd864-0953-44cf-ad93-5284648538f2" (UID: "de9dd864-0953-44cf-ad93-5284648538f2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.085029 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khd78\" (UniqueName: \"kubernetes.io/projected/de9dd864-0953-44cf-ad93-5284648538f2-kube-api-access-khd78\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.085061 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.085074 4904 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de9dd864-0953-44cf-ad93-5284648538f2-logs\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.085085 4904 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.085097 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.085110 4904 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de9dd864-0953-44cf-ad93-5284648538f2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.891187 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.897523 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5","Type":"ContainerStarted","Data":"0ccc1ce88ab841a0e63fd049c6b30208bb6c2ff4d27a39a0abbd9885010f7000"} Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.926566 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.926551655 podStartE2EDuration="2.926551655s" podCreationTimestamp="2026-02-14 11:30:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:30:06.923790649 +0000 UTC m=+1197.736555330" watchObservedRunningTime="2026-02-14 11:30:06.926551655 +0000 UTC m=+1197.739316316" Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.963957 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.970643 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.983688 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 14 11:30:06 crc kubenswrapper[4904]: E0214 11:30:06.984070 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9dd864-0953-44cf-ad93-5284648538f2" containerName="nova-api-api" Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.984089 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9dd864-0953-44cf-ad93-5284648538f2" containerName="nova-api-api" Feb 14 11:30:06 crc kubenswrapper[4904]: E0214 11:30:06.984106 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9dd864-0953-44cf-ad93-5284648538f2" containerName="nova-api-log" Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.984113 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9dd864-0953-44cf-ad93-5284648538f2" containerName="nova-api-log" Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.984274 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="de9dd864-0953-44cf-ad93-5284648538f2" containerName="nova-api-api" Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.984291 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="de9dd864-0953-44cf-ad93-5284648538f2" containerName="nova-api-log" Feb 14 11:30:06 crc kubenswrapper[4904]: I0214 11:30:06.985200 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:06.988825 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:06.988992 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:06.989103 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:06.996234 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.120703 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe2df707-f559-4976-bda3-a0c2b58cb0cf-logs\") pod \"nova-api-0\" (UID: \"fe2df707-f559-4976-bda3-a0c2b58cb0cf\") " pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.120776 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe2df707-f559-4976-bda3-a0c2b58cb0cf-config-data\") pod \"nova-api-0\" (UID: \"fe2df707-f559-4976-bda3-a0c2b58cb0cf\") " pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.120856 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe2df707-f559-4976-bda3-a0c2b58cb0cf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fe2df707-f559-4976-bda3-a0c2b58cb0cf\") " pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.120877 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe2df707-f559-4976-bda3-a0c2b58cb0cf-internal-tls-certs\") pod \"nova-api-0\" (UID: \"fe2df707-f559-4976-bda3-a0c2b58cb0cf\") " pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.121012 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8vcn\" (UniqueName: \"kubernetes.io/projected/fe2df707-f559-4976-bda3-a0c2b58cb0cf-kube-api-access-d8vcn\") pod \"nova-api-0\" (UID: \"fe2df707-f559-4976-bda3-a0c2b58cb0cf\") " pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.121043 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe2df707-f559-4976-bda3-a0c2b58cb0cf-public-tls-certs\") pod \"nova-api-0\" (UID: \"fe2df707-f559-4976-bda3-a0c2b58cb0cf\") " pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.223193 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe2df707-f559-4976-bda3-a0c2b58cb0cf-logs\") pod \"nova-api-0\" (UID: \"fe2df707-f559-4976-bda3-a0c2b58cb0cf\") " pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.223272 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe2df707-f559-4976-bda3-a0c2b58cb0cf-config-data\") pod \"nova-api-0\" (UID: \"fe2df707-f559-4976-bda3-a0c2b58cb0cf\") " pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.223346 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe2df707-f559-4976-bda3-a0c2b58cb0cf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fe2df707-f559-4976-bda3-a0c2b58cb0cf\") " pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.223369 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe2df707-f559-4976-bda3-a0c2b58cb0cf-internal-tls-certs\") pod \"nova-api-0\" (UID: \"fe2df707-f559-4976-bda3-a0c2b58cb0cf\") " pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.223435 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8vcn\" (UniqueName: \"kubernetes.io/projected/fe2df707-f559-4976-bda3-a0c2b58cb0cf-kube-api-access-d8vcn\") pod \"nova-api-0\" (UID: \"fe2df707-f559-4976-bda3-a0c2b58cb0cf\") " pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.223467 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe2df707-f559-4976-bda3-a0c2b58cb0cf-public-tls-certs\") pod \"nova-api-0\" (UID: \"fe2df707-f559-4976-bda3-a0c2b58cb0cf\") " pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.223610 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe2df707-f559-4976-bda3-a0c2b58cb0cf-logs\") pod \"nova-api-0\" (UID: \"fe2df707-f559-4976-bda3-a0c2b58cb0cf\") " pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.227258 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe2df707-f559-4976-bda3-a0c2b58cb0cf-public-tls-certs\") pod \"nova-api-0\" (UID: \"fe2df707-f559-4976-bda3-a0c2b58cb0cf\") " pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.227656 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe2df707-f559-4976-bda3-a0c2b58cb0cf-config-data\") pod \"nova-api-0\" (UID: \"fe2df707-f559-4976-bda3-a0c2b58cb0cf\") " pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.228529 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe2df707-f559-4976-bda3-a0c2b58cb0cf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fe2df707-f559-4976-bda3-a0c2b58cb0cf\") " pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.235795 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe2df707-f559-4976-bda3-a0c2b58cb0cf-internal-tls-certs\") pod \"nova-api-0\" (UID: \"fe2df707-f559-4976-bda3-a0c2b58cb0cf\") " pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.238388 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8vcn\" (UniqueName: \"kubernetes.io/projected/fe2df707-f559-4976-bda3-a0c2b58cb0cf-kube-api-access-d8vcn\") pod \"nova-api-0\" (UID: \"fe2df707-f559-4976-bda3-a0c2b58cb0cf\") " pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.341038 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.774231 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.846748 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de9dd864-0953-44cf-ad93-5284648538f2" path="/var/lib/kubelet/pods/de9dd864-0953-44cf-ad93-5284648538f2/volumes" Feb 14 11:30:07 crc kubenswrapper[4904]: I0214 11:30:07.909826 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fe2df707-f559-4976-bda3-a0c2b58cb0cf","Type":"ContainerStarted","Data":"e32989807af8b776eca65375cc18d6ee2f1375d2bc5129f5cf6c3f71b0f035f6"} Feb 14 11:30:08 crc kubenswrapper[4904]: I0214 11:30:08.542895 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 14 11:30:08 crc kubenswrapper[4904]: I0214 11:30:08.543203 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 14 11:30:08 crc kubenswrapper[4904]: I0214 11:30:08.928892 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fe2df707-f559-4976-bda3-a0c2b58cb0cf","Type":"ContainerStarted","Data":"8be54b9d3b2ab277098b970cb263e5150c90bfd62369e906af231628e9f60379"} Feb 14 11:30:08 crc kubenswrapper[4904]: I0214 11:30:08.928935 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fe2df707-f559-4976-bda3-a0c2b58cb0cf","Type":"ContainerStarted","Data":"61814f9c631041a83a8af48049bdd011cefbd877f15b0073050201a24f6b1b79"} Feb 14 11:30:08 crc kubenswrapper[4904]: I0214 11:30:08.960816 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.960791055 podStartE2EDuration="2.960791055s" podCreationTimestamp="2026-02-14 11:30:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:30:08.954015798 +0000 UTC m=+1199.766780479" watchObservedRunningTime="2026-02-14 11:30:08.960791055 +0000 UTC m=+1199.773555736" Feb 14 11:30:10 crc kubenswrapper[4904]: I0214 11:30:10.267397 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 14 11:30:13 crc kubenswrapper[4904]: I0214 11:30:13.543797 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 14 11:30:13 crc kubenswrapper[4904]: I0214 11:30:13.544193 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 14 11:30:14 crc kubenswrapper[4904]: I0214 11:30:14.556954 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="aba968e8-e711-4cbc-aea2-6e68c7c9a27c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.209:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 14 11:30:14 crc kubenswrapper[4904]: I0214 11:30:14.557004 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="aba968e8-e711-4cbc-aea2-6e68c7c9a27c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.209:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 14 11:30:15 crc kubenswrapper[4904]: I0214 11:30:15.268303 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 14 11:30:15 crc kubenswrapper[4904]: I0214 11:30:15.317555 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 14 11:30:16 crc kubenswrapper[4904]: I0214 11:30:16.023250 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 14 11:30:17 crc kubenswrapper[4904]: I0214 11:30:17.341844 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 14 11:30:17 crc kubenswrapper[4904]: I0214 11:30:17.342136 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 14 11:30:18 crc kubenswrapper[4904]: I0214 11:30:18.356032 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fe2df707-f559-4976-bda3-a0c2b58cb0cf" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.211:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 14 11:30:18 crc kubenswrapper[4904]: I0214 11:30:18.356717 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fe2df707-f559-4976-bda3-a0c2b58cb0cf" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.211:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 14 11:30:19 crc kubenswrapper[4904]: I0214 11:30:19.404101 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 14 11:30:23 crc kubenswrapper[4904]: I0214 11:30:23.552216 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 14 11:30:23 crc kubenswrapper[4904]: I0214 11:30:23.553593 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 14 11:30:23 crc kubenswrapper[4904]: I0214 11:30:23.560487 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 14 11:30:24 crc kubenswrapper[4904]: I0214 11:30:24.095214 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 14 11:30:27 crc kubenswrapper[4904]: I0214 11:30:27.349617 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 14 11:30:27 crc kubenswrapper[4904]: I0214 11:30:27.350286 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 14 11:30:27 crc kubenswrapper[4904]: I0214 11:30:27.350322 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 14 11:30:27 crc kubenswrapper[4904]: I0214 11:30:27.356253 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 14 11:30:28 crc kubenswrapper[4904]: I0214 11:30:28.124682 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 14 11:30:28 crc kubenswrapper[4904]: I0214 11:30:28.132443 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 14 11:30:35 crc kubenswrapper[4904]: I0214 11:30:35.688151 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 14 11:30:36 crc kubenswrapper[4904]: I0214 11:30:36.463584 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 14 11:30:45 crc kubenswrapper[4904]: I0214 11:30:45.940578 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="2b06dd61-da9f-455f-b813-53300e4fbc10" containerName="rabbitmq" containerID="cri-o://10cf7b5b8da00578e784761802e52e4c79c9f67fb379dd1ff190c0d554e6dff3" gracePeriod=604790 Feb 14 11:30:46 crc kubenswrapper[4904]: I0214 11:30:46.182475 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="7448d80b-d340-46a2-88a6-221db1f93f11" containerName="rabbitmq" containerID="cri-o://b2d172a2f150af5def63664c78711f02a0a8bdfe5561905a918dde89ad2f8c0d" gracePeriod=604791 Feb 14 11:30:46 crc kubenswrapper[4904]: I0214 11:30:46.729483 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="2b06dd61-da9f-455f-b813-53300e4fbc10" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.94:5671: connect: connection refused" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.133748 4904 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="7448d80b-d340-46a2-88a6-221db1f93f11" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.95:5671: connect: connection refused" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.522617 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-z77tf"] Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.524476 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.529553 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.544926 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-z77tf"] Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.693799 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.693944 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.693975 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.694000 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.694017 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwdfr\" (UniqueName: \"kubernetes.io/projected/61682878-9897-43bd-a60f-94d30431b829-kube-api-access-bwdfr\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.694069 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.694105 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-config\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.795732 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.796132 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-config\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.796188 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.796268 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.796305 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.796338 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.796365 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwdfr\" (UniqueName: \"kubernetes.io/projected/61682878-9897-43bd-a60f-94d30431b829-kube-api-access-bwdfr\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.797710 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.797725 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.798088 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-config\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.798384 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.798403 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.798663 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.818227 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwdfr\" (UniqueName: \"kubernetes.io/projected/61682878-9897-43bd-a60f-94d30431b829-kube-api-access-bwdfr\") pod \"dnsmasq-dns-79bd4cc8c9-z77tf\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.843186 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:47 crc kubenswrapper[4904]: I0214 11:30:47.973487 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.100194 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7448d80b-d340-46a2-88a6-221db1f93f11-pod-info\") pod \"7448d80b-d340-46a2-88a6-221db1f93f11\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.100316 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7448d80b-d340-46a2-88a6-221db1f93f11-server-conf\") pod \"7448d80b-d340-46a2-88a6-221db1f93f11\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.100372 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgz54\" (UniqueName: \"kubernetes.io/projected/7448d80b-d340-46a2-88a6-221db1f93f11-kube-api-access-zgz54\") pod \"7448d80b-d340-46a2-88a6-221db1f93f11\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.100421 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-tls\") pod \"7448d80b-d340-46a2-88a6-221db1f93f11\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.100456 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7448d80b-d340-46a2-88a6-221db1f93f11-config-data\") pod \"7448d80b-d340-46a2-88a6-221db1f93f11\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.100519 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"7448d80b-d340-46a2-88a6-221db1f93f11\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.100552 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-plugins\") pod \"7448d80b-d340-46a2-88a6-221db1f93f11\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.100582 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7448d80b-d340-46a2-88a6-221db1f93f11-erlang-cookie-secret\") pod \"7448d80b-d340-46a2-88a6-221db1f93f11\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.100606 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7448d80b-d340-46a2-88a6-221db1f93f11-plugins-conf\") pod \"7448d80b-d340-46a2-88a6-221db1f93f11\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.100625 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-confd\") pod \"7448d80b-d340-46a2-88a6-221db1f93f11\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.100732 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-erlang-cookie\") pod \"7448d80b-d340-46a2-88a6-221db1f93f11\" (UID: \"7448d80b-d340-46a2-88a6-221db1f93f11\") " Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.102545 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "7448d80b-d340-46a2-88a6-221db1f93f11" (UID: "7448d80b-d340-46a2-88a6-221db1f93f11"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.107726 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7448d80b-d340-46a2-88a6-221db1f93f11-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "7448d80b-d340-46a2-88a6-221db1f93f11" (UID: "7448d80b-d340-46a2-88a6-221db1f93f11"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.108636 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "7448d80b-d340-46a2-88a6-221db1f93f11" (UID: "7448d80b-d340-46a2-88a6-221db1f93f11"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.109511 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/7448d80b-d340-46a2-88a6-221db1f93f11-pod-info" (OuterVolumeSpecName: "pod-info") pod "7448d80b-d340-46a2-88a6-221db1f93f11" (UID: "7448d80b-d340-46a2-88a6-221db1f93f11"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.112206 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7448d80b-d340-46a2-88a6-221db1f93f11-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "7448d80b-d340-46a2-88a6-221db1f93f11" (UID: "7448d80b-d340-46a2-88a6-221db1f93f11"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.113504 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7448d80b-d340-46a2-88a6-221db1f93f11-kube-api-access-zgz54" (OuterVolumeSpecName: "kube-api-access-zgz54") pod "7448d80b-d340-46a2-88a6-221db1f93f11" (UID: "7448d80b-d340-46a2-88a6-221db1f93f11"). InnerVolumeSpecName "kube-api-access-zgz54". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.114827 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "7448d80b-d340-46a2-88a6-221db1f93f11" (UID: "7448d80b-d340-46a2-88a6-221db1f93f11"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.118938 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "7448d80b-d340-46a2-88a6-221db1f93f11" (UID: "7448d80b-d340-46a2-88a6-221db1f93f11"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.170279 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7448d80b-d340-46a2-88a6-221db1f93f11-config-data" (OuterVolumeSpecName: "config-data") pod "7448d80b-d340-46a2-88a6-221db1f93f11" (UID: "7448d80b-d340-46a2-88a6-221db1f93f11"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.202884 4904 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.202914 4904 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.202923 4904 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7448d80b-d340-46a2-88a6-221db1f93f11-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.202932 4904 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7448d80b-d340-46a2-88a6-221db1f93f11-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.202943 4904 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.202951 4904 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7448d80b-d340-46a2-88a6-221db1f93f11-pod-info\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.202962 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgz54\" (UniqueName: \"kubernetes.io/projected/7448d80b-d340-46a2-88a6-221db1f93f11-kube-api-access-zgz54\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.202970 4904 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.203169 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7448d80b-d340-46a2-88a6-221db1f93f11-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.208071 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7448d80b-d340-46a2-88a6-221db1f93f11-server-conf" (OuterVolumeSpecName: "server-conf") pod "7448d80b-d340-46a2-88a6-221db1f93f11" (UID: "7448d80b-d340-46a2-88a6-221db1f93f11"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.230286 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "7448d80b-d340-46a2-88a6-221db1f93f11" (UID: "7448d80b-d340-46a2-88a6-221db1f93f11"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.230384 4904 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.304301 4904 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7448d80b-d340-46a2-88a6-221db1f93f11-server-conf\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.304332 4904 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.304341 4904 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7448d80b-d340-46a2-88a6-221db1f93f11-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.310502 4904 generic.go:334] "Generic (PLEG): container finished" podID="7448d80b-d340-46a2-88a6-221db1f93f11" containerID="b2d172a2f150af5def63664c78711f02a0a8bdfe5561905a918dde89ad2f8c0d" exitCode=0 Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.310539 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7448d80b-d340-46a2-88a6-221db1f93f11","Type":"ContainerDied","Data":"b2d172a2f150af5def63664c78711f02a0a8bdfe5561905a918dde89ad2f8c0d"} Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.310566 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7448d80b-d340-46a2-88a6-221db1f93f11","Type":"ContainerDied","Data":"1f9964be6c39625296f551e8e5b8f52617d69edd9c1bb0fdc725f9fdfe524226"} Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.310575 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.310585 4904 scope.go:117] "RemoveContainer" containerID="b2d172a2f150af5def63664c78711f02a0a8bdfe5561905a918dde89ad2f8c0d" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.335940 4904 scope.go:117] "RemoveContainer" containerID="4e2f01e425ebbd9b9326491b905f29fd60a6633fb15da5c6854ae3eebc906136" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.345148 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.356408 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.375545 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 14 11:30:48 crc kubenswrapper[4904]: E0214 11:30:48.375900 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7448d80b-d340-46a2-88a6-221db1f93f11" containerName="setup-container" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.375916 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="7448d80b-d340-46a2-88a6-221db1f93f11" containerName="setup-container" Feb 14 11:30:48 crc kubenswrapper[4904]: E0214 11:30:48.375944 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7448d80b-d340-46a2-88a6-221db1f93f11" containerName="rabbitmq" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.375950 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="7448d80b-d340-46a2-88a6-221db1f93f11" containerName="rabbitmq" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.376139 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="7448d80b-d340-46a2-88a6-221db1f93f11" containerName="rabbitmq" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.380478 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.390753 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.390946 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.391061 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.391200 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.411813 4904 scope.go:117] "RemoveContainer" containerID="b2d172a2f150af5def63664c78711f02a0a8bdfe5561905a918dde89ad2f8c0d" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.412157 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.412399 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 14 11:30:48 crc kubenswrapper[4904]: E0214 11:30:48.412397 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2d172a2f150af5def63664c78711f02a0a8bdfe5561905a918dde89ad2f8c0d\": container with ID starting with b2d172a2f150af5def63664c78711f02a0a8bdfe5561905a918dde89ad2f8c0d not found: ID does not exist" containerID="b2d172a2f150af5def63664c78711f02a0a8bdfe5561905a918dde89ad2f8c0d" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.412526 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-6jpz2" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.412608 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2d172a2f150af5def63664c78711f02a0a8bdfe5561905a918dde89ad2f8c0d"} err="failed to get container status \"b2d172a2f150af5def63664c78711f02a0a8bdfe5561905a918dde89ad2f8c0d\": rpc error: code = NotFound desc = could not find container \"b2d172a2f150af5def63664c78711f02a0a8bdfe5561905a918dde89ad2f8c0d\": container with ID starting with b2d172a2f150af5def63664c78711f02a0a8bdfe5561905a918dde89ad2f8c0d not found: ID does not exist" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.412712 4904 scope.go:117] "RemoveContainer" containerID="4e2f01e425ebbd9b9326491b905f29fd60a6633fb15da5c6854ae3eebc906136" Feb 14 11:30:48 crc kubenswrapper[4904]: E0214 11:30:48.413996 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e2f01e425ebbd9b9326491b905f29fd60a6633fb15da5c6854ae3eebc906136\": container with ID starting with 4e2f01e425ebbd9b9326491b905f29fd60a6633fb15da5c6854ae3eebc906136 not found: ID does not exist" containerID="4e2f01e425ebbd9b9326491b905f29fd60a6633fb15da5c6854ae3eebc906136" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.414035 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e2f01e425ebbd9b9326491b905f29fd60a6633fb15da5c6854ae3eebc906136"} err="failed to get container status \"4e2f01e425ebbd9b9326491b905f29fd60a6633fb15da5c6854ae3eebc906136\": rpc error: code = NotFound desc = could not find container \"4e2f01e425ebbd9b9326491b905f29fd60a6633fb15da5c6854ae3eebc906136\": container with ID starting with 4e2f01e425ebbd9b9326491b905f29fd60a6633fb15da5c6854ae3eebc906136 not found: ID does not exist" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.417118 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.463605 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-z77tf"] Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.508016 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/82a277e1-cec5-43a6-a7d4-6651965c02a7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.508051 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/82a277e1-cec5-43a6-a7d4-6651965c02a7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.508119 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82a277e1-cec5-43a6-a7d4-6651965c02a7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.508140 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/82a277e1-cec5-43a6-a7d4-6651965c02a7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.508157 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/82a277e1-cec5-43a6-a7d4-6651965c02a7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.508181 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/82a277e1-cec5-43a6-a7d4-6651965c02a7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.508207 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5dxr\" (UniqueName: \"kubernetes.io/projected/82a277e1-cec5-43a6-a7d4-6651965c02a7-kube-api-access-m5dxr\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.508224 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/82a277e1-cec5-43a6-a7d4-6651965c02a7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.508269 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/82a277e1-cec5-43a6-a7d4-6651965c02a7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.508298 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.508320 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/82a277e1-cec5-43a6-a7d4-6651965c02a7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.610169 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/82a277e1-cec5-43a6-a7d4-6651965c02a7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.610234 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/82a277e1-cec5-43a6-a7d4-6651965c02a7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.610258 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/82a277e1-cec5-43a6-a7d4-6651965c02a7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.611371 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82a277e1-cec5-43a6-a7d4-6651965c02a7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.611408 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/82a277e1-cec5-43a6-a7d4-6651965c02a7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.611437 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/82a277e1-cec5-43a6-a7d4-6651965c02a7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.611470 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/82a277e1-cec5-43a6-a7d4-6651965c02a7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.611510 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5dxr\" (UniqueName: \"kubernetes.io/projected/82a277e1-cec5-43a6-a7d4-6651965c02a7-kube-api-access-m5dxr\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.611537 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/82a277e1-cec5-43a6-a7d4-6651965c02a7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.611610 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/82a277e1-cec5-43a6-a7d4-6651965c02a7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.611657 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.611771 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/82a277e1-cec5-43a6-a7d4-6651965c02a7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.611952 4904 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.612237 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/82a277e1-cec5-43a6-a7d4-6651965c02a7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.612481 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/82a277e1-cec5-43a6-a7d4-6651965c02a7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.612820 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82a277e1-cec5-43a6-a7d4-6651965c02a7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.612858 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/82a277e1-cec5-43a6-a7d4-6651965c02a7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.617023 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/82a277e1-cec5-43a6-a7d4-6651965c02a7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.618509 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/82a277e1-cec5-43a6-a7d4-6651965c02a7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.619485 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/82a277e1-cec5-43a6-a7d4-6651965c02a7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.619492 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/82a277e1-cec5-43a6-a7d4-6651965c02a7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.631294 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5dxr\" (UniqueName: \"kubernetes.io/projected/82a277e1-cec5-43a6-a7d4-6651965c02a7-kube-api-access-m5dxr\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.671324 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"82a277e1-cec5-43a6-a7d4-6651965c02a7\") " pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:48 crc kubenswrapper[4904]: I0214 11:30:48.934598 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:30:49 crc kubenswrapper[4904]: I0214 11:30:49.321749 4904 generic.go:334] "Generic (PLEG): container finished" podID="61682878-9897-43bd-a60f-94d30431b829" containerID="aff661ca39174cafc98ab9e8330349d97e88a72507155fd8b30747dd679fbba3" exitCode=0 Feb 14 11:30:49 crc kubenswrapper[4904]: I0214 11:30:49.321946 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" event={"ID":"61682878-9897-43bd-a60f-94d30431b829","Type":"ContainerDied","Data":"aff661ca39174cafc98ab9e8330349d97e88a72507155fd8b30747dd679fbba3"} Feb 14 11:30:49 crc kubenswrapper[4904]: I0214 11:30:49.322127 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" event={"ID":"61682878-9897-43bd-a60f-94d30431b829","Type":"ContainerStarted","Data":"d3a5a9319c2a4cae82efd9e33677773bdca453b942c494e3eadc050711c3a48d"} Feb 14 11:30:49 crc kubenswrapper[4904]: I0214 11:30:49.501385 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 14 11:30:49 crc kubenswrapper[4904]: W0214 11:30:49.514059 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82a277e1_cec5_43a6_a7d4_6651965c02a7.slice/crio-33a4f5ace97f077e9b6129f8e2b952d15627fb93898f7c20e9a8cfe6abe944cc WatchSource:0}: Error finding container 33a4f5ace97f077e9b6129f8e2b952d15627fb93898f7c20e9a8cfe6abe944cc: Status 404 returned error can't find the container with id 33a4f5ace97f077e9b6129f8e2b952d15627fb93898f7c20e9a8cfe6abe944cc Feb 14 11:30:49 crc kubenswrapper[4904]: I0214 11:30:49.874610 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7448d80b-d340-46a2-88a6-221db1f93f11" path="/var/lib/kubelet/pods/7448d80b-d340-46a2-88a6-221db1f93f11/volumes" Feb 14 11:30:50 crc kubenswrapper[4904]: I0214 11:30:50.338791 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"82a277e1-cec5-43a6-a7d4-6651965c02a7","Type":"ContainerStarted","Data":"33a4f5ace97f077e9b6129f8e2b952d15627fb93898f7c20e9a8cfe6abe944cc"} Feb 14 11:30:50 crc kubenswrapper[4904]: I0214 11:30:50.340973 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" event={"ID":"61682878-9897-43bd-a60f-94d30431b829","Type":"ContainerStarted","Data":"04d9eaaa31fdb43a149e5ed059dc34cbfd997ee601addf20f6fc35883e5c6568"} Feb 14 11:30:50 crc kubenswrapper[4904]: I0214 11:30:50.342960 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:51 crc kubenswrapper[4904]: I0214 11:30:51.357485 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"82a277e1-cec5-43a6-a7d4-6651965c02a7","Type":"ContainerStarted","Data":"f74a1ded93e0a885fd044cb7406c9c5ca6febfc7d55db2b6772e1f0b3f386259"} Feb 14 11:30:51 crc kubenswrapper[4904]: I0214 11:30:51.383908 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" podStartSLOduration=4.383889445 podStartE2EDuration="4.383889445s" podCreationTimestamp="2026-02-14 11:30:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:30:50.359787202 +0000 UTC m=+1241.172551873" watchObservedRunningTime="2026-02-14 11:30:51.383889445 +0000 UTC m=+1242.196654116" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.367006 4904 generic.go:334] "Generic (PLEG): container finished" podID="2b06dd61-da9f-455f-b813-53300e4fbc10" containerID="10cf7b5b8da00578e784761802e52e4c79c9f67fb379dd1ff190c0d554e6dff3" exitCode=0 Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.367084 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2b06dd61-da9f-455f-b813-53300e4fbc10","Type":"ContainerDied","Data":"10cf7b5b8da00578e784761802e52e4c79c9f67fb379dd1ff190c0d554e6dff3"} Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.648269 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.705310 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2b06dd61-da9f-455f-b813-53300e4fbc10-config-data\") pod \"2b06dd61-da9f-455f-b813-53300e4fbc10\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.705392 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"2b06dd61-da9f-455f-b813-53300e4fbc10\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.705435 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-confd\") pod \"2b06dd61-da9f-455f-b813-53300e4fbc10\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.705493 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ctwr\" (UniqueName: \"kubernetes.io/projected/2b06dd61-da9f-455f-b813-53300e4fbc10-kube-api-access-9ctwr\") pod \"2b06dd61-da9f-455f-b813-53300e4fbc10\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.705511 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2b06dd61-da9f-455f-b813-53300e4fbc10-pod-info\") pod \"2b06dd61-da9f-455f-b813-53300e4fbc10\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.705545 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2b06dd61-da9f-455f-b813-53300e4fbc10-server-conf\") pod \"2b06dd61-da9f-455f-b813-53300e4fbc10\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.705598 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-tls\") pod \"2b06dd61-da9f-455f-b813-53300e4fbc10\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.705622 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2b06dd61-da9f-455f-b813-53300e4fbc10-plugins-conf\") pod \"2b06dd61-da9f-455f-b813-53300e4fbc10\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.705698 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-plugins\") pod \"2b06dd61-da9f-455f-b813-53300e4fbc10\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.705732 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-erlang-cookie\") pod \"2b06dd61-da9f-455f-b813-53300e4fbc10\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.705754 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2b06dd61-da9f-455f-b813-53300e4fbc10-erlang-cookie-secret\") pod \"2b06dd61-da9f-455f-b813-53300e4fbc10\" (UID: \"2b06dd61-da9f-455f-b813-53300e4fbc10\") " Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.708155 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "2b06dd61-da9f-455f-b813-53300e4fbc10" (UID: "2b06dd61-da9f-455f-b813-53300e4fbc10"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.708491 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b06dd61-da9f-455f-b813-53300e4fbc10-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "2b06dd61-da9f-455f-b813-53300e4fbc10" (UID: "2b06dd61-da9f-455f-b813-53300e4fbc10"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.710178 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "2b06dd61-da9f-455f-b813-53300e4fbc10" (UID: "2b06dd61-da9f-455f-b813-53300e4fbc10"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.719027 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/2b06dd61-da9f-455f-b813-53300e4fbc10-pod-info" (OuterVolumeSpecName: "pod-info") pod "2b06dd61-da9f-455f-b813-53300e4fbc10" (UID: "2b06dd61-da9f-455f-b813-53300e4fbc10"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.721336 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "2b06dd61-da9f-455f-b813-53300e4fbc10" (UID: "2b06dd61-da9f-455f-b813-53300e4fbc10"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.736345 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b06dd61-da9f-455f-b813-53300e4fbc10-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "2b06dd61-da9f-455f-b813-53300e4fbc10" (UID: "2b06dd61-da9f-455f-b813-53300e4fbc10"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.759922 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "2b06dd61-da9f-455f-b813-53300e4fbc10" (UID: "2b06dd61-da9f-455f-b813-53300e4fbc10"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.760369 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b06dd61-da9f-455f-b813-53300e4fbc10-kube-api-access-9ctwr" (OuterVolumeSpecName: "kube-api-access-9ctwr") pod "2b06dd61-da9f-455f-b813-53300e4fbc10" (UID: "2b06dd61-da9f-455f-b813-53300e4fbc10"). InnerVolumeSpecName "kube-api-access-9ctwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.783406 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b06dd61-da9f-455f-b813-53300e4fbc10-config-data" (OuterVolumeSpecName: "config-data") pod "2b06dd61-da9f-455f-b813-53300e4fbc10" (UID: "2b06dd61-da9f-455f-b813-53300e4fbc10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.805535 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b06dd61-da9f-455f-b813-53300e4fbc10-server-conf" (OuterVolumeSpecName: "server-conf") pod "2b06dd61-da9f-455f-b813-53300e4fbc10" (UID: "2b06dd61-da9f-455f-b813-53300e4fbc10"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.808273 4904 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2b06dd61-da9f-455f-b813-53300e4fbc10-pod-info\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.808293 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ctwr\" (UniqueName: \"kubernetes.io/projected/2b06dd61-da9f-455f-b813-53300e4fbc10-kube-api-access-9ctwr\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.808304 4904 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2b06dd61-da9f-455f-b813-53300e4fbc10-server-conf\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.808323 4904 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.808331 4904 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2b06dd61-da9f-455f-b813-53300e4fbc10-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.808338 4904 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.808346 4904 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.808353 4904 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2b06dd61-da9f-455f-b813-53300e4fbc10-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.808361 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2b06dd61-da9f-455f-b813-53300e4fbc10-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.808386 4904 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.830952 4904 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.903761 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "2b06dd61-da9f-455f-b813-53300e4fbc10" (UID: "2b06dd61-da9f-455f-b813-53300e4fbc10"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.910403 4904 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:52 crc kubenswrapper[4904]: I0214 11:30:52.910462 4904 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2b06dd61-da9f-455f-b813-53300e4fbc10-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.375249 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2b06dd61-da9f-455f-b813-53300e4fbc10","Type":"ContainerDied","Data":"f1209a24707701731254d181030bc253c4623d294926a0f93e47a725aa24a711"} Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.375523 4904 scope.go:117] "RemoveContainer" containerID="10cf7b5b8da00578e784761802e52e4c79c9f67fb379dd1ff190c0d554e6dff3" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.375290 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.634569 4904 scope.go:117] "RemoveContainer" containerID="67f4cf2e5ed352d8d50d8a7b06dbd4929f867edd15e88f56f90d69d36544af7b" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.711687 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.739198 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.743388 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 14 11:30:53 crc kubenswrapper[4904]: E0214 11:30:53.743898 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b06dd61-da9f-455f-b813-53300e4fbc10" containerName="rabbitmq" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.747198 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b06dd61-da9f-455f-b813-53300e4fbc10" containerName="rabbitmq" Feb 14 11:30:53 crc kubenswrapper[4904]: E0214 11:30:53.747280 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b06dd61-da9f-455f-b813-53300e4fbc10" containerName="setup-container" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.747333 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b06dd61-da9f-455f-b813-53300e4fbc10" containerName="setup-container" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.747597 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b06dd61-da9f-455f-b813-53300e4fbc10" containerName="rabbitmq" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.754056 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.757436 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.765851 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.766073 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.766198 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.766299 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.766717 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.767257 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.772928 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-9qlhw" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.823672 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/87e6f193-7618-42af-a3b3-107cbda62c46-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.823725 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/87e6f193-7618-42af-a3b3-107cbda62c46-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.823807 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/87e6f193-7618-42af-a3b3-107cbda62c46-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.823871 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/87e6f193-7618-42af-a3b3-107cbda62c46-pod-info\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.823900 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/87e6f193-7618-42af-a3b3-107cbda62c46-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.823926 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/87e6f193-7618-42af-a3b3-107cbda62c46-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.823956 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/87e6f193-7618-42af-a3b3-107cbda62c46-server-conf\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.823974 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.824039 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/87e6f193-7618-42af-a3b3-107cbda62c46-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.824062 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjsh9\" (UniqueName: \"kubernetes.io/projected/87e6f193-7618-42af-a3b3-107cbda62c46-kube-api-access-bjsh9\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.824086 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/87e6f193-7618-42af-a3b3-107cbda62c46-config-data\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.851369 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b06dd61-da9f-455f-b813-53300e4fbc10" path="/var/lib/kubelet/pods/2b06dd61-da9f-455f-b813-53300e4fbc10/volumes" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.924984 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/87e6f193-7618-42af-a3b3-107cbda62c46-pod-info\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.925072 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/87e6f193-7618-42af-a3b3-107cbda62c46-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.925154 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/87e6f193-7618-42af-a3b3-107cbda62c46-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.925642 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/87e6f193-7618-42af-a3b3-107cbda62c46-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.925686 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/87e6f193-7618-42af-a3b3-107cbda62c46-server-conf\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.925708 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.926438 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/87e6f193-7618-42af-a3b3-107cbda62c46-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.926465 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjsh9\" (UniqueName: \"kubernetes.io/projected/87e6f193-7618-42af-a3b3-107cbda62c46-kube-api-access-bjsh9\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.926492 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/87e6f193-7618-42af-a3b3-107cbda62c46-config-data\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.926510 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/87e6f193-7618-42af-a3b3-107cbda62c46-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.926532 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/87e6f193-7618-42af-a3b3-107cbda62c46-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.926623 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/87e6f193-7618-42af-a3b3-107cbda62c46-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.926945 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/87e6f193-7618-42af-a3b3-107cbda62c46-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.926365 4904 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.927337 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/87e6f193-7618-42af-a3b3-107cbda62c46-server-conf\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.929418 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/87e6f193-7618-42af-a3b3-107cbda62c46-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.932322 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/87e6f193-7618-42af-a3b3-107cbda62c46-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.932775 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/87e6f193-7618-42af-a3b3-107cbda62c46-config-data\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.935041 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/87e6f193-7618-42af-a3b3-107cbda62c46-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.935598 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/87e6f193-7618-42af-a3b3-107cbda62c46-pod-info\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.938689 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/87e6f193-7618-42af-a3b3-107cbda62c46-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.967321 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjsh9\" (UniqueName: \"kubernetes.io/projected/87e6f193-7618-42af-a3b3-107cbda62c46-kube-api-access-bjsh9\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:53 crc kubenswrapper[4904]: I0214 11:30:53.969870 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"87e6f193-7618-42af-a3b3-107cbda62c46\") " pod="openstack/rabbitmq-server-0" Feb 14 11:30:54 crc kubenswrapper[4904]: I0214 11:30:54.083520 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 14 11:30:54 crc kubenswrapper[4904]: I0214 11:30:54.580116 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 14 11:30:55 crc kubenswrapper[4904]: I0214 11:30:55.393944 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"87e6f193-7618-42af-a3b3-107cbda62c46","Type":"ContainerStarted","Data":"3d3f8c8397072a35dbc67bf1eeeafee59aa79211457e253534cb8cf8b711ee28"} Feb 14 11:30:56 crc kubenswrapper[4904]: I0214 11:30:56.405162 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"87e6f193-7618-42af-a3b3-107cbda62c46","Type":"ContainerStarted","Data":"4ebb23b9cf55e0fc72c6c08c929f3be6d1362fb7712f18f39df069553550f1fe"} Feb 14 11:30:57 crc kubenswrapper[4904]: I0214 11:30:57.848410 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:30:57 crc kubenswrapper[4904]: I0214 11:30:57.912193 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-bl2fg"] Feb 14 11:30:57 crc kubenswrapper[4904]: I0214 11:30:57.912735 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" podUID="0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41" containerName="dnsmasq-dns" containerID="cri-o://90052437023c3c0bd09b40b8d6216c637bcdc4e70e68e5a12b619632522d303c" gracePeriod=10 Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.083396 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d7677974f-mz7hn"] Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.085072 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.113055 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d7677974f-mz7hn"] Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.208068 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fa3f3a5c-2d8d-424c-a590-e101c89443a4-openstack-edpm-ipam\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.208145 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa3f3a5c-2d8d-424c-a590-e101c89443a4-ovsdbserver-nb\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.208169 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa3f3a5c-2d8d-424c-a590-e101c89443a4-config\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.208375 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxnb9\" (UniqueName: \"kubernetes.io/projected/fa3f3a5c-2d8d-424c-a590-e101c89443a4-kube-api-access-mxnb9\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.208448 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fa3f3a5c-2d8d-424c-a590-e101c89443a4-ovsdbserver-sb\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.208509 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa3f3a5c-2d8d-424c-a590-e101c89443a4-dns-svc\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.208622 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fa3f3a5c-2d8d-424c-a590-e101c89443a4-dns-swift-storage-0\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.310781 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fa3f3a5c-2d8d-424c-a590-e101c89443a4-openstack-edpm-ipam\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.310829 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa3f3a5c-2d8d-424c-a590-e101c89443a4-ovsdbserver-nb\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.310861 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa3f3a5c-2d8d-424c-a590-e101c89443a4-config\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.310923 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxnb9\" (UniqueName: \"kubernetes.io/projected/fa3f3a5c-2d8d-424c-a590-e101c89443a4-kube-api-access-mxnb9\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.310952 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fa3f3a5c-2d8d-424c-a590-e101c89443a4-ovsdbserver-sb\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.310982 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa3f3a5c-2d8d-424c-a590-e101c89443a4-dns-svc\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.311014 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fa3f3a5c-2d8d-424c-a590-e101c89443a4-dns-swift-storage-0\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.312416 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fa3f3a5c-2d8d-424c-a590-e101c89443a4-dns-swift-storage-0\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.312535 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa3f3a5c-2d8d-424c-a590-e101c89443a4-config\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.312550 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fa3f3a5c-2d8d-424c-a590-e101c89443a4-openstack-edpm-ipam\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.313349 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa3f3a5c-2d8d-424c-a590-e101c89443a4-ovsdbserver-nb\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.313413 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fa3f3a5c-2d8d-424c-a590-e101c89443a4-ovsdbserver-sb\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.314487 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa3f3a5c-2d8d-424c-a590-e101c89443a4-dns-svc\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.333258 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxnb9\" (UniqueName: \"kubernetes.io/projected/fa3f3a5c-2d8d-424c-a590-e101c89443a4-kube-api-access-mxnb9\") pod \"dnsmasq-dns-d7677974f-mz7hn\" (UID: \"fa3f3a5c-2d8d-424c-a590-e101c89443a4\") " pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.433951 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.437610 4904 generic.go:334] "Generic (PLEG): container finished" podID="0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41" containerID="90052437023c3c0bd09b40b8d6216c637bcdc4e70e68e5a12b619632522d303c" exitCode=0 Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.437648 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" event={"ID":"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41","Type":"ContainerDied","Data":"90052437023c3c0bd09b40b8d6216c637bcdc4e70e68e5a12b619632522d303c"} Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.437674 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" event={"ID":"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41","Type":"ContainerDied","Data":"87f077076dde91c45574e80c731fffaa26eaeee68e929d83b7eb7f6ed139797d"} Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.437683 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87f077076dde91c45574e80c731fffaa26eaeee68e929d83b7eb7f6ed139797d" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.522179 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.614634 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-dns-svc\") pod \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.614704 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-config\") pod \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.614808 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-ovsdbserver-sb\") pod \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.614893 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-dns-swift-storage-0\") pod \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.614951 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-ovsdbserver-nb\") pod \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.615013 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxjfk\" (UniqueName: \"kubernetes.io/projected/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-kube-api-access-rxjfk\") pod \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\" (UID: \"0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41\") " Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.619254 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-kube-api-access-rxjfk" (OuterVolumeSpecName: "kube-api-access-rxjfk") pod "0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41" (UID: "0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41"). InnerVolumeSpecName "kube-api-access-rxjfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.675713 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-config" (OuterVolumeSpecName: "config") pod "0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41" (UID: "0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.680507 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41" (UID: "0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.698266 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41" (UID: "0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.712567 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41" (UID: "0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.717512 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxjfk\" (UniqueName: \"kubernetes.io/projected/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-kube-api-access-rxjfk\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.717541 4904 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.717553 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.717564 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.717574 4904 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.718073 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41" (UID: "0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.818739 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 14 11:30:58 crc kubenswrapper[4904]: W0214 11:30:58.934828 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa3f3a5c_2d8d_424c_a590_e101c89443a4.slice/crio-0278f5b1d5424f7a4b5eff2cafd5e0340900fcf59c59900853d7f8d9e5afa5d0 WatchSource:0}: Error finding container 0278f5b1d5424f7a4b5eff2cafd5e0340900fcf59c59900853d7f8d9e5afa5d0: Status 404 returned error can't find the container with id 0278f5b1d5424f7a4b5eff2cafd5e0340900fcf59c59900853d7f8d9e5afa5d0 Feb 14 11:30:58 crc kubenswrapper[4904]: I0214 11:30:58.936510 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d7677974f-mz7hn"] Feb 14 11:30:59 crc kubenswrapper[4904]: I0214 11:30:59.449776 4904 generic.go:334] "Generic (PLEG): container finished" podID="fa3f3a5c-2d8d-424c-a590-e101c89443a4" containerID="3aa8be034e29381489d4234d9ef3ff4363e3db13f737585fc70c93dcba31c3d0" exitCode=0 Feb 14 11:30:59 crc kubenswrapper[4904]: I0214 11:30:59.449890 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d7677974f-mz7hn" event={"ID":"fa3f3a5c-2d8d-424c-a590-e101c89443a4","Type":"ContainerDied","Data":"3aa8be034e29381489d4234d9ef3ff4363e3db13f737585fc70c93dcba31c3d0"} Feb 14 11:30:59 crc kubenswrapper[4904]: I0214 11:30:59.450352 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d7677974f-mz7hn" event={"ID":"fa3f3a5c-2d8d-424c-a590-e101c89443a4","Type":"ContainerStarted","Data":"0278f5b1d5424f7a4b5eff2cafd5e0340900fcf59c59900853d7f8d9e5afa5d0"} Feb 14 11:30:59 crc kubenswrapper[4904]: I0214 11:30:59.450468 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-bl2fg" Feb 14 11:30:59 crc kubenswrapper[4904]: I0214 11:30:59.670484 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-bl2fg"] Feb 14 11:30:59 crc kubenswrapper[4904]: I0214 11:30:59.692498 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-bl2fg"] Feb 14 11:30:59 crc kubenswrapper[4904]: I0214 11:30:59.859689 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41" path="/var/lib/kubelet/pods/0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41/volumes" Feb 14 11:31:00 crc kubenswrapper[4904]: I0214 11:31:00.459540 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d7677974f-mz7hn" event={"ID":"fa3f3a5c-2d8d-424c-a590-e101c89443a4","Type":"ContainerStarted","Data":"19be6cdc8b1e86771047b9d987d8cbdb7a32ac4a13e53618874e9cf8d565da96"} Feb 14 11:31:00 crc kubenswrapper[4904]: I0214 11:31:00.459950 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:31:00 crc kubenswrapper[4904]: I0214 11:31:00.486393 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d7677974f-mz7hn" podStartSLOduration=2.486370477 podStartE2EDuration="2.486370477s" podCreationTimestamp="2026-02-14 11:30:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:31:00.480212197 +0000 UTC m=+1251.292976868" watchObservedRunningTime="2026-02-14 11:31:00.486370477 +0000 UTC m=+1251.299135138" Feb 14 11:31:08 crc kubenswrapper[4904]: I0214 11:31:08.435812 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d7677974f-mz7hn" Feb 14 11:31:08 crc kubenswrapper[4904]: I0214 11:31:08.512487 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-z77tf"] Feb 14 11:31:08 crc kubenswrapper[4904]: I0214 11:31:08.512776 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" podUID="61682878-9897-43bd-a60f-94d30431b829" containerName="dnsmasq-dns" containerID="cri-o://04d9eaaa31fdb43a149e5ed059dc34cbfd997ee601addf20f6fc35883e5c6568" gracePeriod=10 Feb 14 11:31:08 crc kubenswrapper[4904]: I0214 11:31:08.998453 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.108387 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-config\") pod \"61682878-9897-43bd-a60f-94d30431b829\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.108480 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwdfr\" (UniqueName: \"kubernetes.io/projected/61682878-9897-43bd-a60f-94d30431b829-kube-api-access-bwdfr\") pod \"61682878-9897-43bd-a60f-94d30431b829\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.108510 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-openstack-edpm-ipam\") pod \"61682878-9897-43bd-a60f-94d30431b829\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.108541 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-dns-svc\") pod \"61682878-9897-43bd-a60f-94d30431b829\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.108641 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-dns-swift-storage-0\") pod \"61682878-9897-43bd-a60f-94d30431b829\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.108734 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-ovsdbserver-nb\") pod \"61682878-9897-43bd-a60f-94d30431b829\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.108778 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-ovsdbserver-sb\") pod \"61682878-9897-43bd-a60f-94d30431b829\" (UID: \"61682878-9897-43bd-a60f-94d30431b829\") " Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.165507 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61682878-9897-43bd-a60f-94d30431b829-kube-api-access-bwdfr" (OuterVolumeSpecName: "kube-api-access-bwdfr") pod "61682878-9897-43bd-a60f-94d30431b829" (UID: "61682878-9897-43bd-a60f-94d30431b829"). InnerVolumeSpecName "kube-api-access-bwdfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.212997 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwdfr\" (UniqueName: \"kubernetes.io/projected/61682878-9897-43bd-a60f-94d30431b829-kube-api-access-bwdfr\") on node \"crc\" DevicePath \"\"" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.259896 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "61682878-9897-43bd-a60f-94d30431b829" (UID: "61682878-9897-43bd-a60f-94d30431b829"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.266856 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "61682878-9897-43bd-a60f-94d30431b829" (UID: "61682878-9897-43bd-a60f-94d30431b829"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.277814 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-config" (OuterVolumeSpecName: "config") pod "61682878-9897-43bd-a60f-94d30431b829" (UID: "61682878-9897-43bd-a60f-94d30431b829"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.282250 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "61682878-9897-43bd-a60f-94d30431b829" (UID: "61682878-9897-43bd-a60f-94d30431b829"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.285285 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "61682878-9897-43bd-a60f-94d30431b829" (UID: "61682878-9897-43bd-a60f-94d30431b829"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.291610 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "61682878-9897-43bd-a60f-94d30431b829" (UID: "61682878-9897-43bd-a60f-94d30431b829"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.314364 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.314392 4904 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.314401 4904 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-config\") on node \"crc\" DevicePath \"\"" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.314411 4904 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.314421 4904 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.314429 4904 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/61682878-9897-43bd-a60f-94d30431b829-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.551082 4904 generic.go:334] "Generic (PLEG): container finished" podID="61682878-9897-43bd-a60f-94d30431b829" containerID="04d9eaaa31fdb43a149e5ed059dc34cbfd997ee601addf20f6fc35883e5c6568" exitCode=0 Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.551451 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" event={"ID":"61682878-9897-43bd-a60f-94d30431b829","Type":"ContainerDied","Data":"04d9eaaa31fdb43a149e5ed059dc34cbfd997ee601addf20f6fc35883e5c6568"} Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.551508 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" event={"ID":"61682878-9897-43bd-a60f-94d30431b829","Type":"ContainerDied","Data":"d3a5a9319c2a4cae82efd9e33677773bdca453b942c494e3eadc050711c3a48d"} Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.551531 4904 scope.go:117] "RemoveContainer" containerID="04d9eaaa31fdb43a149e5ed059dc34cbfd997ee601addf20f6fc35883e5c6568" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.551826 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-z77tf" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.577079 4904 scope.go:117] "RemoveContainer" containerID="aff661ca39174cafc98ab9e8330349d97e88a72507155fd8b30747dd679fbba3" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.601303 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-z77tf"] Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.611411 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-z77tf"] Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.627299 4904 scope.go:117] "RemoveContainer" containerID="04d9eaaa31fdb43a149e5ed059dc34cbfd997ee601addf20f6fc35883e5c6568" Feb 14 11:31:09 crc kubenswrapper[4904]: E0214 11:31:09.627783 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04d9eaaa31fdb43a149e5ed059dc34cbfd997ee601addf20f6fc35883e5c6568\": container with ID starting with 04d9eaaa31fdb43a149e5ed059dc34cbfd997ee601addf20f6fc35883e5c6568 not found: ID does not exist" containerID="04d9eaaa31fdb43a149e5ed059dc34cbfd997ee601addf20f6fc35883e5c6568" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.627942 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04d9eaaa31fdb43a149e5ed059dc34cbfd997ee601addf20f6fc35883e5c6568"} err="failed to get container status \"04d9eaaa31fdb43a149e5ed059dc34cbfd997ee601addf20f6fc35883e5c6568\": rpc error: code = NotFound desc = could not find container \"04d9eaaa31fdb43a149e5ed059dc34cbfd997ee601addf20f6fc35883e5c6568\": container with ID starting with 04d9eaaa31fdb43a149e5ed059dc34cbfd997ee601addf20f6fc35883e5c6568 not found: ID does not exist" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.628091 4904 scope.go:117] "RemoveContainer" containerID="aff661ca39174cafc98ab9e8330349d97e88a72507155fd8b30747dd679fbba3" Feb 14 11:31:09 crc kubenswrapper[4904]: E0214 11:31:09.628503 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aff661ca39174cafc98ab9e8330349d97e88a72507155fd8b30747dd679fbba3\": container with ID starting with aff661ca39174cafc98ab9e8330349d97e88a72507155fd8b30747dd679fbba3 not found: ID does not exist" containerID="aff661ca39174cafc98ab9e8330349d97e88a72507155fd8b30747dd679fbba3" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.628609 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aff661ca39174cafc98ab9e8330349d97e88a72507155fd8b30747dd679fbba3"} err="failed to get container status \"aff661ca39174cafc98ab9e8330349d97e88a72507155fd8b30747dd679fbba3\": rpc error: code = NotFound desc = could not find container \"aff661ca39174cafc98ab9e8330349d97e88a72507155fd8b30747dd679fbba3\": container with ID starting with aff661ca39174cafc98ab9e8330349d97e88a72507155fd8b30747dd679fbba3 not found: ID does not exist" Feb 14 11:31:09 crc kubenswrapper[4904]: I0214 11:31:09.848666 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61682878-9897-43bd-a60f-94d30431b829" path="/var/lib/kubelet/pods/61682878-9897-43bd-a60f-94d30431b829/volumes" Feb 14 11:31:23 crc kubenswrapper[4904]: I0214 11:31:23.688336 4904 generic.go:334] "Generic (PLEG): container finished" podID="82a277e1-cec5-43a6-a7d4-6651965c02a7" containerID="f74a1ded93e0a885fd044cb7406c9c5ca6febfc7d55db2b6772e1f0b3f386259" exitCode=0 Feb 14 11:31:23 crc kubenswrapper[4904]: I0214 11:31:23.688418 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"82a277e1-cec5-43a6-a7d4-6651965c02a7","Type":"ContainerDied","Data":"f74a1ded93e0a885fd044cb7406c9c5ca6febfc7d55db2b6772e1f0b3f386259"} Feb 14 11:31:24 crc kubenswrapper[4904]: I0214 11:31:24.697177 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"82a277e1-cec5-43a6-a7d4-6651965c02a7","Type":"ContainerStarted","Data":"bb369a2431c8d5cf2cb7d17ccc782ce1c86f07f275d91a74188052fbf44cae47"} Feb 14 11:31:24 crc kubenswrapper[4904]: I0214 11:31:24.698478 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:31:24 crc kubenswrapper[4904]: I0214 11:31:24.723309 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.723291728 podStartE2EDuration="36.723291728s" podCreationTimestamp="2026-02-14 11:30:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:31:24.716351597 +0000 UTC m=+1275.529116258" watchObservedRunningTime="2026-02-14 11:31:24.723291728 +0000 UTC m=+1275.536056379" Feb 14 11:31:28 crc kubenswrapper[4904]: I0214 11:31:28.729798 4904 generic.go:334] "Generic (PLEG): container finished" podID="87e6f193-7618-42af-a3b3-107cbda62c46" containerID="4ebb23b9cf55e0fc72c6c08c929f3be6d1362fb7712f18f39df069553550f1fe" exitCode=0 Feb 14 11:31:28 crc kubenswrapper[4904]: I0214 11:31:28.729872 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"87e6f193-7618-42af-a3b3-107cbda62c46","Type":"ContainerDied","Data":"4ebb23b9cf55e0fc72c6c08c929f3be6d1362fb7712f18f39df069553550f1fe"} Feb 14 11:31:29 crc kubenswrapper[4904]: I0214 11:31:29.745526 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"87e6f193-7618-42af-a3b3-107cbda62c46","Type":"ContainerStarted","Data":"b72148e19decd3b4ca01cfd4f57ac90d0f558efa44422ae78fe5d40a3adf5141"} Feb 14 11:31:29 crc kubenswrapper[4904]: I0214 11:31:29.748121 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 14 11:31:29 crc kubenswrapper[4904]: I0214 11:31:29.783326 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.783310106 podStartE2EDuration="36.783310106s" podCreationTimestamp="2026-02-14 11:30:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 11:31:29.77764576 +0000 UTC m=+1280.590410431" watchObservedRunningTime="2026-02-14 11:31:29.783310106 +0000 UTC m=+1280.596074767" Feb 14 11:31:31 crc kubenswrapper[4904]: I0214 11:31:31.911028 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z"] Feb 14 11:31:31 crc kubenswrapper[4904]: E0214 11:31:31.911711 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61682878-9897-43bd-a60f-94d30431b829" containerName="dnsmasq-dns" Feb 14 11:31:31 crc kubenswrapper[4904]: I0214 11:31:31.911722 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="61682878-9897-43bd-a60f-94d30431b829" containerName="dnsmasq-dns" Feb 14 11:31:31 crc kubenswrapper[4904]: E0214 11:31:31.911750 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41" containerName="init" Feb 14 11:31:31 crc kubenswrapper[4904]: I0214 11:31:31.911756 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41" containerName="init" Feb 14 11:31:31 crc kubenswrapper[4904]: E0214 11:31:31.911768 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41" containerName="dnsmasq-dns" Feb 14 11:31:31 crc kubenswrapper[4904]: I0214 11:31:31.911774 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41" containerName="dnsmasq-dns" Feb 14 11:31:31 crc kubenswrapper[4904]: E0214 11:31:31.911789 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61682878-9897-43bd-a60f-94d30431b829" containerName="init" Feb 14 11:31:31 crc kubenswrapper[4904]: I0214 11:31:31.911794 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="61682878-9897-43bd-a60f-94d30431b829" containerName="init" Feb 14 11:31:31 crc kubenswrapper[4904]: I0214 11:31:31.911972 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="61682878-9897-43bd-a60f-94d30431b829" containerName="dnsmasq-dns" Feb 14 11:31:31 crc kubenswrapper[4904]: I0214 11:31:31.911991 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a70ccdc-1b57-4ecd-ad1c-20fc67c41f41" containerName="dnsmasq-dns" Feb 14 11:31:31 crc kubenswrapper[4904]: I0214 11:31:31.912571 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" Feb 14 11:31:31 crc kubenswrapper[4904]: I0214 11:31:31.914638 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-48rnz" Feb 14 11:31:31 crc kubenswrapper[4904]: I0214 11:31:31.914825 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 14 11:31:31 crc kubenswrapper[4904]: I0214 11:31:31.914965 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 14 11:31:31 crc kubenswrapper[4904]: I0214 11:31:31.916682 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 14 11:31:31 crc kubenswrapper[4904]: I0214 11:31:31.946178 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z"] Feb 14 11:31:31 crc kubenswrapper[4904]: I0214 11:31:31.972357 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2mvx\" (UniqueName: \"kubernetes.io/projected/245410d0-e53c-4d88-8c47-369c93f449c5-kube-api-access-h2mvx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z\" (UID: \"245410d0-e53c-4d88-8c47-369c93f449c5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" Feb 14 11:31:31 crc kubenswrapper[4904]: I0214 11:31:31.972505 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/245410d0-e53c-4d88-8c47-369c93f449c5-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z\" (UID: \"245410d0-e53c-4d88-8c47-369c93f449c5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" Feb 14 11:31:31 crc kubenswrapper[4904]: I0214 11:31:31.972596 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/245410d0-e53c-4d88-8c47-369c93f449c5-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z\" (UID: \"245410d0-e53c-4d88-8c47-369c93f449c5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" Feb 14 11:31:31 crc kubenswrapper[4904]: I0214 11:31:31.972649 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/245410d0-e53c-4d88-8c47-369c93f449c5-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z\" (UID: \"245410d0-e53c-4d88-8c47-369c93f449c5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" Feb 14 11:31:32 crc kubenswrapper[4904]: I0214 11:31:32.074132 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/245410d0-e53c-4d88-8c47-369c93f449c5-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z\" (UID: \"245410d0-e53c-4d88-8c47-369c93f449c5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" Feb 14 11:31:32 crc kubenswrapper[4904]: I0214 11:31:32.074210 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/245410d0-e53c-4d88-8c47-369c93f449c5-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z\" (UID: \"245410d0-e53c-4d88-8c47-369c93f449c5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" Feb 14 11:31:32 crc kubenswrapper[4904]: I0214 11:31:32.074252 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/245410d0-e53c-4d88-8c47-369c93f449c5-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z\" (UID: \"245410d0-e53c-4d88-8c47-369c93f449c5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" Feb 14 11:31:32 crc kubenswrapper[4904]: I0214 11:31:32.074299 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2mvx\" (UniqueName: \"kubernetes.io/projected/245410d0-e53c-4d88-8c47-369c93f449c5-kube-api-access-h2mvx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z\" (UID: \"245410d0-e53c-4d88-8c47-369c93f449c5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" Feb 14 11:31:32 crc kubenswrapper[4904]: I0214 11:31:32.080544 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/245410d0-e53c-4d88-8c47-369c93f449c5-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z\" (UID: \"245410d0-e53c-4d88-8c47-369c93f449c5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" Feb 14 11:31:32 crc kubenswrapper[4904]: I0214 11:31:32.081116 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/245410d0-e53c-4d88-8c47-369c93f449c5-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z\" (UID: \"245410d0-e53c-4d88-8c47-369c93f449c5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" Feb 14 11:31:32 crc kubenswrapper[4904]: I0214 11:31:32.081382 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/245410d0-e53c-4d88-8c47-369c93f449c5-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z\" (UID: \"245410d0-e53c-4d88-8c47-369c93f449c5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" Feb 14 11:31:32 crc kubenswrapper[4904]: I0214 11:31:32.103202 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2mvx\" (UniqueName: \"kubernetes.io/projected/245410d0-e53c-4d88-8c47-369c93f449c5-kube-api-access-h2mvx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z\" (UID: \"245410d0-e53c-4d88-8c47-369c93f449c5\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" Feb 14 11:31:32 crc kubenswrapper[4904]: I0214 11:31:32.230257 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" Feb 14 11:31:32 crc kubenswrapper[4904]: I0214 11:31:32.976894 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z"] Feb 14 11:31:33 crc kubenswrapper[4904]: I0214 11:31:33.779225 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" event={"ID":"245410d0-e53c-4d88-8c47-369c93f449c5","Type":"ContainerStarted","Data":"355ce532150ea04e16475a5dbcc4c3a0de00012c2a968c485d39e44d03d9f5e8"} Feb 14 11:31:38 crc kubenswrapper[4904]: I0214 11:31:38.940052 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 14 11:31:42 crc kubenswrapper[4904]: I0214 11:31:42.866973 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" event={"ID":"245410d0-e53c-4d88-8c47-369c93f449c5","Type":"ContainerStarted","Data":"2427418e297949b7587bd0f743cf52c64b54741465b014d5dc679244a61e9dca"} Feb 14 11:31:42 crc kubenswrapper[4904]: I0214 11:31:42.883527 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" podStartSLOduration=2.476438097 podStartE2EDuration="11.883513094s" podCreationTimestamp="2026-02-14 11:31:31 +0000 UTC" firstStartedPulling="2026-02-14 11:31:32.9861653 +0000 UTC m=+1283.798929961" lastFinishedPulling="2026-02-14 11:31:42.393240267 +0000 UTC m=+1293.206004958" observedRunningTime="2026-02-14 11:31:42.882056134 +0000 UTC m=+1293.694820795" watchObservedRunningTime="2026-02-14 11:31:42.883513094 +0000 UTC m=+1293.696277755" Feb 14 11:31:44 crc kubenswrapper[4904]: I0214 11:31:44.088041 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 14 11:31:46 crc kubenswrapper[4904]: I0214 11:31:46.383199 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:31:46 crc kubenswrapper[4904]: I0214 11:31:46.383481 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:32:13 crc kubenswrapper[4904]: I0214 11:32:13.391053 4904 scope.go:117] "RemoveContainer" containerID="2b926ba49a0893e1bd8e39df58f1784fd89df5b8d794db995e453a4fdc763e98" Feb 14 11:32:13 crc kubenswrapper[4904]: I0214 11:32:13.439633 4904 scope.go:117] "RemoveContainer" containerID="cabf8e56e34a6df0b7dde1428f2b80975722c3e1dc55210f459dd51cd7ab17a7" Feb 14 11:32:13 crc kubenswrapper[4904]: I0214 11:32:13.468765 4904 scope.go:117] "RemoveContainer" containerID="9f345d8012a3bbd34736da7a1d8efa129d14b0c2092b8099c1cb6addc6a8c5ca" Feb 14 11:32:13 crc kubenswrapper[4904]: I0214 11:32:13.487520 4904 scope.go:117] "RemoveContainer" containerID="a2103801ef86e1b867adb135f0a9513174a6c5cc3638429966de7427d29c4333" Feb 14 11:32:13 crc kubenswrapper[4904]: I0214 11:32:13.534631 4904 scope.go:117] "RemoveContainer" containerID="522c8e0477ec7ed39163c6dc7ca5c2542ed209a4caa41b362756bc571a8ad301" Feb 14 11:32:13 crc kubenswrapper[4904]: I0214 11:32:13.576476 4904 scope.go:117] "RemoveContainer" containerID="0d7a1d3ca53e5ca1de252c08d0b8846188358ec24880dd6a1b0426d8befd8763" Feb 14 11:32:13 crc kubenswrapper[4904]: I0214 11:32:13.597222 4904 scope.go:117] "RemoveContainer" containerID="f7b2c2d2959ca38c222d69f4400b8387898daab9f22ebbc77c383e5b4d87cb3b" Feb 14 11:32:16 crc kubenswrapper[4904]: I0214 11:32:16.382576 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:32:16 crc kubenswrapper[4904]: I0214 11:32:16.382920 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:32:46 crc kubenswrapper[4904]: I0214 11:32:46.383031 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:32:46 crc kubenswrapper[4904]: I0214 11:32:46.383702 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:32:46 crc kubenswrapper[4904]: I0214 11:32:46.383772 4904 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:32:46 crc kubenswrapper[4904]: I0214 11:32:46.385038 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7753982cea734b21f061129ef8b3fab0b493e6bfec1d88e26788a0cb6fd15b70"} pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 14 11:32:46 crc kubenswrapper[4904]: I0214 11:32:46.385351 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" containerID="cri-o://7753982cea734b21f061129ef8b3fab0b493e6bfec1d88e26788a0cb6fd15b70" gracePeriod=600 Feb 14 11:32:46 crc kubenswrapper[4904]: I0214 11:32:46.524723 4904 generic.go:334] "Generic (PLEG): container finished" podID="7790d276-2506-4641-8c22-67e6441417f0" containerID="7753982cea734b21f061129ef8b3fab0b493e6bfec1d88e26788a0cb6fd15b70" exitCode=0 Feb 14 11:32:46 crc kubenswrapper[4904]: I0214 11:32:46.524776 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerDied","Data":"7753982cea734b21f061129ef8b3fab0b493e6bfec1d88e26788a0cb6fd15b70"} Feb 14 11:32:46 crc kubenswrapper[4904]: I0214 11:32:46.524864 4904 scope.go:117] "RemoveContainer" containerID="d567bf42576edbcc9fce256f34faf5a41303db84f6e9f300d9f0592cbb6a48e2" Feb 14 11:32:47 crc kubenswrapper[4904]: I0214 11:32:47.550497 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc"} Feb 14 11:33:13 crc kubenswrapper[4904]: I0214 11:33:13.734670 4904 scope.go:117] "RemoveContainer" containerID="6bb8ce5e37d9faa7fb04dbc186646c0baff6a994d8e2887ed71781d05f2a5b2d" Feb 14 11:33:59 crc kubenswrapper[4904]: I0214 11:33:59.185979 4904 generic.go:334] "Generic (PLEG): container finished" podID="245410d0-e53c-4d88-8c47-369c93f449c5" containerID="2427418e297949b7587bd0f743cf52c64b54741465b014d5dc679244a61e9dca" exitCode=2 Feb 14 11:33:59 crc kubenswrapper[4904]: I0214 11:33:59.186071 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" event={"ID":"245410d0-e53c-4d88-8c47-369c93f449c5","Type":"ContainerDied","Data":"2427418e297949b7587bd0f743cf52c64b54741465b014d5dc679244a61e9dca"} Feb 14 11:34:00 crc kubenswrapper[4904]: I0214 11:34:00.642067 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" Feb 14 11:34:00 crc kubenswrapper[4904]: I0214 11:34:00.769359 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2mvx\" (UniqueName: \"kubernetes.io/projected/245410d0-e53c-4d88-8c47-369c93f449c5-kube-api-access-h2mvx\") pod \"245410d0-e53c-4d88-8c47-369c93f449c5\" (UID: \"245410d0-e53c-4d88-8c47-369c93f449c5\") " Feb 14 11:34:00 crc kubenswrapper[4904]: I0214 11:34:00.769455 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/245410d0-e53c-4d88-8c47-369c93f449c5-inventory\") pod \"245410d0-e53c-4d88-8c47-369c93f449c5\" (UID: \"245410d0-e53c-4d88-8c47-369c93f449c5\") " Feb 14 11:34:00 crc kubenswrapper[4904]: I0214 11:34:00.769514 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/245410d0-e53c-4d88-8c47-369c93f449c5-repo-setup-combined-ca-bundle\") pod \"245410d0-e53c-4d88-8c47-369c93f449c5\" (UID: \"245410d0-e53c-4d88-8c47-369c93f449c5\") " Feb 14 11:34:00 crc kubenswrapper[4904]: I0214 11:34:00.769548 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/245410d0-e53c-4d88-8c47-369c93f449c5-ssh-key-openstack-edpm-ipam\") pod \"245410d0-e53c-4d88-8c47-369c93f449c5\" (UID: \"245410d0-e53c-4d88-8c47-369c93f449c5\") " Feb 14 11:34:00 crc kubenswrapper[4904]: I0214 11:34:00.775044 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/245410d0-e53c-4d88-8c47-369c93f449c5-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "245410d0-e53c-4d88-8c47-369c93f449c5" (UID: "245410d0-e53c-4d88-8c47-369c93f449c5"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:34:00 crc kubenswrapper[4904]: I0214 11:34:00.784290 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/245410d0-e53c-4d88-8c47-369c93f449c5-kube-api-access-h2mvx" (OuterVolumeSpecName: "kube-api-access-h2mvx") pod "245410d0-e53c-4d88-8c47-369c93f449c5" (UID: "245410d0-e53c-4d88-8c47-369c93f449c5"). InnerVolumeSpecName "kube-api-access-h2mvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:34:00 crc kubenswrapper[4904]: I0214 11:34:00.797823 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/245410d0-e53c-4d88-8c47-369c93f449c5-inventory" (OuterVolumeSpecName: "inventory") pod "245410d0-e53c-4d88-8c47-369c93f449c5" (UID: "245410d0-e53c-4d88-8c47-369c93f449c5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:34:00 crc kubenswrapper[4904]: I0214 11:34:00.800961 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/245410d0-e53c-4d88-8c47-369c93f449c5-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "245410d0-e53c-4d88-8c47-369c93f449c5" (UID: "245410d0-e53c-4d88-8c47-369c93f449c5"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:34:00 crc kubenswrapper[4904]: I0214 11:34:00.872028 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2mvx\" (UniqueName: \"kubernetes.io/projected/245410d0-e53c-4d88-8c47-369c93f449c5-kube-api-access-h2mvx\") on node \"crc\" DevicePath \"\"" Feb 14 11:34:00 crc kubenswrapper[4904]: I0214 11:34:00.872064 4904 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/245410d0-e53c-4d88-8c47-369c93f449c5-inventory\") on node \"crc\" DevicePath \"\"" Feb 14 11:34:00 crc kubenswrapper[4904]: I0214 11:34:00.872075 4904 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/245410d0-e53c-4d88-8c47-369c93f449c5-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:34:00 crc kubenswrapper[4904]: I0214 11:34:00.872111 4904 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/245410d0-e53c-4d88-8c47-369c93f449c5-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 14 11:34:01 crc kubenswrapper[4904]: I0214 11:34:01.209295 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" event={"ID":"245410d0-e53c-4d88-8c47-369c93f449c5","Type":"ContainerDied","Data":"355ce532150ea04e16475a5dbcc4c3a0de00012c2a968c485d39e44d03d9f5e8"} Feb 14 11:34:01 crc kubenswrapper[4904]: I0214 11:34:01.209340 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="355ce532150ea04e16475a5dbcc4c3a0de00012c2a968c485d39e44d03d9f5e8" Feb 14 11:34:01 crc kubenswrapper[4904]: I0214 11:34:01.209400 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.032958 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl"] Feb 14 11:34:08 crc kubenswrapper[4904]: E0214 11:34:08.033997 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="245410d0-e53c-4d88-8c47-369c93f449c5" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.034019 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="245410d0-e53c-4d88-8c47-369c93f449c5" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.034277 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="245410d0-e53c-4d88-8c47-369c93f449c5" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.035122 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.042069 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.042217 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.042463 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.042617 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-48rnz" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.053006 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl"] Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.106149 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w72p\" (UniqueName: \"kubernetes.io/projected/9243ea89-2a6d-43a8-85d7-b83b3e664680-kube-api-access-7w72p\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl\" (UID: \"9243ea89-2a6d-43a8-85d7-b83b3e664680\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.106329 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9243ea89-2a6d-43a8-85d7-b83b3e664680-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl\" (UID: \"9243ea89-2a6d-43a8-85d7-b83b3e664680\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.106572 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9243ea89-2a6d-43a8-85d7-b83b3e664680-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl\" (UID: \"9243ea89-2a6d-43a8-85d7-b83b3e664680\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.106693 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9243ea89-2a6d-43a8-85d7-b83b3e664680-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl\" (UID: \"9243ea89-2a6d-43a8-85d7-b83b3e664680\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.208519 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9243ea89-2a6d-43a8-85d7-b83b3e664680-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl\" (UID: \"9243ea89-2a6d-43a8-85d7-b83b3e664680\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.208591 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w72p\" (UniqueName: \"kubernetes.io/projected/9243ea89-2a6d-43a8-85d7-b83b3e664680-kube-api-access-7w72p\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl\" (UID: \"9243ea89-2a6d-43a8-85d7-b83b3e664680\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.208683 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9243ea89-2a6d-43a8-85d7-b83b3e664680-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl\" (UID: \"9243ea89-2a6d-43a8-85d7-b83b3e664680\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.208738 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9243ea89-2a6d-43a8-85d7-b83b3e664680-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl\" (UID: \"9243ea89-2a6d-43a8-85d7-b83b3e664680\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.214435 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9243ea89-2a6d-43a8-85d7-b83b3e664680-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl\" (UID: \"9243ea89-2a6d-43a8-85d7-b83b3e664680\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.216679 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9243ea89-2a6d-43a8-85d7-b83b3e664680-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl\" (UID: \"9243ea89-2a6d-43a8-85d7-b83b3e664680\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.218317 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9243ea89-2a6d-43a8-85d7-b83b3e664680-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl\" (UID: \"9243ea89-2a6d-43a8-85d7-b83b3e664680\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.228218 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w72p\" (UniqueName: \"kubernetes.io/projected/9243ea89-2a6d-43a8-85d7-b83b3e664680-kube-api-access-7w72p\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl\" (UID: \"9243ea89-2a6d-43a8-85d7-b83b3e664680\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.361668 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" Feb 14 11:34:08 crc kubenswrapper[4904]: I0214 11:34:08.892629 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl"] Feb 14 11:34:09 crc kubenswrapper[4904]: I0214 11:34:09.272554 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" event={"ID":"9243ea89-2a6d-43a8-85d7-b83b3e664680","Type":"ContainerStarted","Data":"b38711e376ec0dc52976dec50dcce1552948b13bf91d1d5a5a275eed0734d85c"} Feb 14 11:34:10 crc kubenswrapper[4904]: I0214 11:34:10.290346 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" event={"ID":"9243ea89-2a6d-43a8-85d7-b83b3e664680","Type":"ContainerStarted","Data":"706d7cba65d2770ae9a902837bf00639c6d088f28479a2019eb21028d9e241e8"} Feb 14 11:34:46 crc kubenswrapper[4904]: I0214 11:34:46.382634 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:34:46 crc kubenswrapper[4904]: I0214 11:34:46.383158 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:35:13 crc kubenswrapper[4904]: I0214 11:35:13.818125 4904 scope.go:117] "RemoveContainer" containerID="993c53a048dccfdb8658894e792ceda74488c299bcef3ba041508c31ec4d5811" Feb 14 11:35:13 crc kubenswrapper[4904]: I0214 11:35:13.844936 4904 scope.go:117] "RemoveContainer" containerID="b91611bb9dd7fa92d5828db01ef7d9cd8c52c856b2261816e48f3a3fc7beafb5" Feb 14 11:35:13 crc kubenswrapper[4904]: I0214 11:35:13.888045 4904 scope.go:117] "RemoveContainer" containerID="9a2f9587f1979ffd4ae82561cf6db276a1697a1301b62c00fe16374d977ec528" Feb 14 11:35:13 crc kubenswrapper[4904]: I0214 11:35:13.905571 4904 scope.go:117] "RemoveContainer" containerID="3768c88de79f10e6f25623269a486e6bdc1d014e2e9057fc89fd5446fef3e8ac" Feb 14 11:35:13 crc kubenswrapper[4904]: I0214 11:35:13.926155 4904 scope.go:117] "RemoveContainer" containerID="7a0c207bba00f8b4d5ef34c28c4dca24db9bfb64ea4d7998e5a8bc1f503d0e8d" Feb 14 11:35:13 crc kubenswrapper[4904]: I0214 11:35:13.944361 4904 scope.go:117] "RemoveContainer" containerID="7b09413ebf4351f5c41e8b4116308fc5e094dcb5844973cadd3b4af6776bd444" Feb 14 11:35:13 crc kubenswrapper[4904]: I0214 11:35:13.965206 4904 scope.go:117] "RemoveContainer" containerID="50638b6242c7853e0f6bca5a74f696a49430021d28c350795c3e2ef5f986a091" Feb 14 11:35:16 crc kubenswrapper[4904]: I0214 11:35:16.383111 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:35:16 crc kubenswrapper[4904]: I0214 11:35:16.383608 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:35:44 crc kubenswrapper[4904]: I0214 11:35:44.056485 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" podStartSLOduration=95.646813088 podStartE2EDuration="1m36.056463893s" podCreationTimestamp="2026-02-14 11:34:08 +0000 UTC" firstStartedPulling="2026-02-14 11:34:08.900059992 +0000 UTC m=+1439.712824653" lastFinishedPulling="2026-02-14 11:34:09.309710797 +0000 UTC m=+1440.122475458" observedRunningTime="2026-02-14 11:34:10.306314674 +0000 UTC m=+1441.119079335" watchObservedRunningTime="2026-02-14 11:35:44.056463893 +0000 UTC m=+1534.869228554" Feb 14 11:35:44 crc kubenswrapper[4904]: I0214 11:35:44.067257 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-981b-account-create-update-rt8h8"] Feb 14 11:35:44 crc kubenswrapper[4904]: I0214 11:35:44.083548 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-4dkbs"] Feb 14 11:35:44 crc kubenswrapper[4904]: I0214 11:35:44.095735 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-4dkbs"] Feb 14 11:35:44 crc kubenswrapper[4904]: I0214 11:35:44.107193 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-981b-account-create-update-rt8h8"] Feb 14 11:35:45 crc kubenswrapper[4904]: I0214 11:35:45.031000 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-836b-account-create-update-kcjv5"] Feb 14 11:35:45 crc kubenswrapper[4904]: I0214 11:35:45.040686 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-9cs22"] Feb 14 11:35:45 crc kubenswrapper[4904]: I0214 11:35:45.051080 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-836b-account-create-update-kcjv5"] Feb 14 11:35:45 crc kubenswrapper[4904]: I0214 11:35:45.061228 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-9cs22"] Feb 14 11:35:45 crc kubenswrapper[4904]: I0214 11:35:45.847461 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12e4e520-e219-4768-b5db-5ccadfee8beb" path="/var/lib/kubelet/pods/12e4e520-e219-4768-b5db-5ccadfee8beb/volumes" Feb 14 11:35:45 crc kubenswrapper[4904]: I0214 11:35:45.848304 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="405e6001-9b9c-46bd-8d84-d5005de91285" path="/var/lib/kubelet/pods/405e6001-9b9c-46bd-8d84-d5005de91285/volumes" Feb 14 11:35:45 crc kubenswrapper[4904]: I0214 11:35:45.848857 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8649450d-1384-4d27-a4bb-501ebee69f01" path="/var/lib/kubelet/pods/8649450d-1384-4d27-a4bb-501ebee69f01/volumes" Feb 14 11:35:45 crc kubenswrapper[4904]: I0214 11:35:45.849392 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd9c7534-8ff2-4898-ab20-660140c54857" path="/var/lib/kubelet/pods/cd9c7534-8ff2-4898-ab20-660140c54857/volumes" Feb 14 11:35:46 crc kubenswrapper[4904]: I0214 11:35:46.064763 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-9z2kq"] Feb 14 11:35:46 crc kubenswrapper[4904]: I0214 11:35:46.076153 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-f0fd-account-create-update-kmmtd"] Feb 14 11:35:46 crc kubenswrapper[4904]: I0214 11:35:46.084039 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-9z2kq"] Feb 14 11:35:46 crc kubenswrapper[4904]: I0214 11:35:46.092020 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-f0fd-account-create-update-kmmtd"] Feb 14 11:35:46 crc kubenswrapper[4904]: I0214 11:35:46.383034 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:35:46 crc kubenswrapper[4904]: I0214 11:35:46.383103 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:35:46 crc kubenswrapper[4904]: I0214 11:35:46.383156 4904 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:35:46 crc kubenswrapper[4904]: I0214 11:35:46.384061 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc"} pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 14 11:35:46 crc kubenswrapper[4904]: I0214 11:35:46.384140 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" containerID="cri-o://6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" gracePeriod=600 Feb 14 11:35:46 crc kubenswrapper[4904]: E0214 11:35:46.507980 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:35:47 crc kubenswrapper[4904]: I0214 11:35:47.146911 4904 generic.go:334] "Generic (PLEG): container finished" podID="7790d276-2506-4641-8c22-67e6441417f0" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" exitCode=0 Feb 14 11:35:47 crc kubenswrapper[4904]: I0214 11:35:47.146951 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerDied","Data":"6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc"} Feb 14 11:35:47 crc kubenswrapper[4904]: I0214 11:35:47.147263 4904 scope.go:117] "RemoveContainer" containerID="7753982cea734b21f061129ef8b3fab0b493e6bfec1d88e26788a0cb6fd15b70" Feb 14 11:35:47 crc kubenswrapper[4904]: I0214 11:35:47.148527 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:35:47 crc kubenswrapper[4904]: E0214 11:35:47.148847 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:35:47 crc kubenswrapper[4904]: I0214 11:35:47.846478 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71d5f892-67e3-4854-bee6-1586d712fe8e" path="/var/lib/kubelet/pods/71d5f892-67e3-4854-bee6-1586d712fe8e/volumes" Feb 14 11:35:47 crc kubenswrapper[4904]: I0214 11:35:47.847265 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90945752-c3c4-4aaf-9680-4c5fe89ab1de" path="/var/lib/kubelet/pods/90945752-c3c4-4aaf-9680-4c5fe89ab1de/volumes" Feb 14 11:35:58 crc kubenswrapper[4904]: I0214 11:35:58.837277 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:35:58 crc kubenswrapper[4904]: E0214 11:35:58.838269 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:36:09 crc kubenswrapper[4904]: I0214 11:36:09.843536 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:36:09 crc kubenswrapper[4904]: E0214 11:36:09.844351 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:36:10 crc kubenswrapper[4904]: I0214 11:36:10.030431 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-c4294"] Feb 14 11:36:10 crc kubenswrapper[4904]: I0214 11:36:10.038804 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-c4294"] Feb 14 11:36:11 crc kubenswrapper[4904]: I0214 11:36:11.852155 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73518149-3e47-413e-8bf1-c29279ee89b7" path="/var/lib/kubelet/pods/73518149-3e47-413e-8bf1-c29279ee89b7/volumes" Feb 14 11:36:13 crc kubenswrapper[4904]: I0214 11:36:13.041239 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-8b70-account-create-update-j2fhg"] Feb 14 11:36:13 crc kubenswrapper[4904]: I0214 11:36:13.051747 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-8b70-account-create-update-j2fhg"] Feb 14 11:36:13 crc kubenswrapper[4904]: I0214 11:36:13.062732 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-9ldqj"] Feb 14 11:36:13 crc kubenswrapper[4904]: I0214 11:36:13.073973 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-vfmnn"] Feb 14 11:36:13 crc kubenswrapper[4904]: I0214 11:36:13.083289 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-8fad-account-create-update-k8rxj"] Feb 14 11:36:13 crc kubenswrapper[4904]: I0214 11:36:13.091812 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-8fad-account-create-update-k8rxj"] Feb 14 11:36:13 crc kubenswrapper[4904]: I0214 11:36:13.100394 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-vfmnn"] Feb 14 11:36:13 crc kubenswrapper[4904]: I0214 11:36:13.108043 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-9ldqj"] Feb 14 11:36:13 crc kubenswrapper[4904]: I0214 11:36:13.850863 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4168b6f1-2a3a-410e-a0c6-42783873ab95" path="/var/lib/kubelet/pods/4168b6f1-2a3a-410e-a0c6-42783873ab95/volumes" Feb 14 11:36:13 crc kubenswrapper[4904]: I0214 11:36:13.851784 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9215ece4-8699-4e9a-aac3-1c1d2b6822c2" path="/var/lib/kubelet/pods/9215ece4-8699-4e9a-aac3-1c1d2b6822c2/volumes" Feb 14 11:36:13 crc kubenswrapper[4904]: I0214 11:36:13.852521 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a628e9c6-8ff3-41bb-92ae-69d703f224da" path="/var/lib/kubelet/pods/a628e9c6-8ff3-41bb-92ae-69d703f224da/volumes" Feb 14 11:36:13 crc kubenswrapper[4904]: I0214 11:36:13.853203 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e148cac5-499e-488a-b06b-9aedaa6f2b41" path="/var/lib/kubelet/pods/e148cac5-499e-488a-b06b-9aedaa6f2b41/volumes" Feb 14 11:36:14 crc kubenswrapper[4904]: I0214 11:36:14.025552 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-7m2cz"] Feb 14 11:36:14 crc kubenswrapper[4904]: I0214 11:36:14.034730 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-7m2cz"] Feb 14 11:36:14 crc kubenswrapper[4904]: I0214 11:36:14.045287 4904 scope.go:117] "RemoveContainer" containerID="60584e082f5e434f71f5e7be8e1dec7daffc3a3759a146c30a5a4f18560e78ab" Feb 14 11:36:14 crc kubenswrapper[4904]: I0214 11:36:14.065987 4904 scope.go:117] "RemoveContainer" containerID="6f5ffacb5f290b1aab1e4dbef99e1d5b80188276c4a1162896b14f1d0342ea14" Feb 14 11:36:14 crc kubenswrapper[4904]: I0214 11:36:14.125875 4904 scope.go:117] "RemoveContainer" containerID="fccd7d513406f1a26ed650c2f745b826f5bf5bdbb5a92907e09cddee575b8f43" Feb 14 11:36:14 crc kubenswrapper[4904]: I0214 11:36:14.161110 4904 scope.go:117] "RemoveContainer" containerID="9216cc8fb1514cda0b04dcef5ed09bc4f644ac9951297556f53bceb35fa299f1" Feb 14 11:36:14 crc kubenswrapper[4904]: I0214 11:36:14.206482 4904 scope.go:117] "RemoveContainer" containerID="90052437023c3c0bd09b40b8d6216c637bcdc4e70e68e5a12b619632522d303c" Feb 14 11:36:14 crc kubenswrapper[4904]: I0214 11:36:14.249367 4904 scope.go:117] "RemoveContainer" containerID="21efbd0cae01bcce998ef4d3fd9ca96dc2a196aa780cf0507087acc47a3ad4b1" Feb 14 11:36:14 crc kubenswrapper[4904]: I0214 11:36:14.270571 4904 scope.go:117] "RemoveContainer" containerID="f93e1d601a859b953290f628d69fdf5b1e1f08ec46d991cff6a9d8e5e65ca6b4" Feb 14 11:36:14 crc kubenswrapper[4904]: I0214 11:36:14.315298 4904 scope.go:117] "RemoveContainer" containerID="cf77dce6159e3dfeb40683ce7ba91a167973488c00ea7c742cd4fbad2ffb935c" Feb 14 11:36:14 crc kubenswrapper[4904]: I0214 11:36:14.337086 4904 scope.go:117] "RemoveContainer" containerID="33cbf2577b882f6c3e5c3df634cb48a4602198a44f88474756c48719932734a3" Feb 14 11:36:14 crc kubenswrapper[4904]: I0214 11:36:14.363266 4904 scope.go:117] "RemoveContainer" containerID="6446be203b1275df99c9c26e3aef8f97bbcab08aadbc5021232ba082c4fa7b66" Feb 14 11:36:14 crc kubenswrapper[4904]: I0214 11:36:14.387968 4904 scope.go:117] "RemoveContainer" containerID="2ffffe14db50a91755915b873223a8c2218e7fb45e12f8371cf4ed51af92b2d7" Feb 14 11:36:14 crc kubenswrapper[4904]: I0214 11:36:14.408172 4904 scope.go:117] "RemoveContainer" containerID="ef630af5206762bfa2b64e289409530532689ee5af2db0e5b75aabe5596cef65" Feb 14 11:36:14 crc kubenswrapper[4904]: I0214 11:36:14.427726 4904 scope.go:117] "RemoveContainer" containerID="a283d1c0c7fcfd0e80f44f71d01133f59c16226f7b3998b7e82e7675ce19bb8f" Feb 14 11:36:15 crc kubenswrapper[4904]: I0214 11:36:15.034665 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-dbdb-account-create-update-6drwj"] Feb 14 11:36:15 crc kubenswrapper[4904]: I0214 11:36:15.041575 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-dbdb-account-create-update-6drwj"] Feb 14 11:36:15 crc kubenswrapper[4904]: I0214 11:36:15.846327 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10c49b28-99e8-4e52-9950-bfec689cc77c" path="/var/lib/kubelet/pods/10c49b28-99e8-4e52-9950-bfec689cc77c/volumes" Feb 14 11:36:15 crc kubenswrapper[4904]: I0214 11:36:15.847651 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60a86662-8426-41d7-8eab-bd0ed8ff9fda" path="/var/lib/kubelet/pods/60a86662-8426-41d7-8eab-bd0ed8ff9fda/volumes" Feb 14 11:36:21 crc kubenswrapper[4904]: I0214 11:36:21.837228 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:36:21 crc kubenswrapper[4904]: E0214 11:36:21.837907 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:36:22 crc kubenswrapper[4904]: I0214 11:36:22.451792 4904 generic.go:334] "Generic (PLEG): container finished" podID="9243ea89-2a6d-43a8-85d7-b83b3e664680" containerID="706d7cba65d2770ae9a902837bf00639c6d088f28479a2019eb21028d9e241e8" exitCode=2 Feb 14 11:36:22 crc kubenswrapper[4904]: I0214 11:36:22.451830 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" event={"ID":"9243ea89-2a6d-43a8-85d7-b83b3e664680","Type":"ContainerDied","Data":"706d7cba65d2770ae9a902837bf00639c6d088f28479a2019eb21028d9e241e8"} Feb 14 11:36:23 crc kubenswrapper[4904]: I0214 11:36:23.894532 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" Feb 14 11:36:24 crc kubenswrapper[4904]: I0214 11:36:24.040442 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7w72p\" (UniqueName: \"kubernetes.io/projected/9243ea89-2a6d-43a8-85d7-b83b3e664680-kube-api-access-7w72p\") pod \"9243ea89-2a6d-43a8-85d7-b83b3e664680\" (UID: \"9243ea89-2a6d-43a8-85d7-b83b3e664680\") " Feb 14 11:36:24 crc kubenswrapper[4904]: I0214 11:36:24.040484 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9243ea89-2a6d-43a8-85d7-b83b3e664680-ssh-key-openstack-edpm-ipam\") pod \"9243ea89-2a6d-43a8-85d7-b83b3e664680\" (UID: \"9243ea89-2a6d-43a8-85d7-b83b3e664680\") " Feb 14 11:36:24 crc kubenswrapper[4904]: I0214 11:36:24.040624 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9243ea89-2a6d-43a8-85d7-b83b3e664680-repo-setup-combined-ca-bundle\") pod \"9243ea89-2a6d-43a8-85d7-b83b3e664680\" (UID: \"9243ea89-2a6d-43a8-85d7-b83b3e664680\") " Feb 14 11:36:24 crc kubenswrapper[4904]: I0214 11:36:24.040732 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9243ea89-2a6d-43a8-85d7-b83b3e664680-inventory\") pod \"9243ea89-2a6d-43a8-85d7-b83b3e664680\" (UID: \"9243ea89-2a6d-43a8-85d7-b83b3e664680\") " Feb 14 11:36:24 crc kubenswrapper[4904]: I0214 11:36:24.044774 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-j4tgk"] Feb 14 11:36:24 crc kubenswrapper[4904]: I0214 11:36:24.054662 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-j4tgk"] Feb 14 11:36:24 crc kubenswrapper[4904]: I0214 11:36:24.060214 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9243ea89-2a6d-43a8-85d7-b83b3e664680-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "9243ea89-2a6d-43a8-85d7-b83b3e664680" (UID: "9243ea89-2a6d-43a8-85d7-b83b3e664680"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:36:24 crc kubenswrapper[4904]: I0214 11:36:24.064243 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9243ea89-2a6d-43a8-85d7-b83b3e664680-kube-api-access-7w72p" (OuterVolumeSpecName: "kube-api-access-7w72p") pod "9243ea89-2a6d-43a8-85d7-b83b3e664680" (UID: "9243ea89-2a6d-43a8-85d7-b83b3e664680"). InnerVolumeSpecName "kube-api-access-7w72p". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:36:24 crc kubenswrapper[4904]: I0214 11:36:24.080104 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9243ea89-2a6d-43a8-85d7-b83b3e664680-inventory" (OuterVolumeSpecName: "inventory") pod "9243ea89-2a6d-43a8-85d7-b83b3e664680" (UID: "9243ea89-2a6d-43a8-85d7-b83b3e664680"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:36:24 crc kubenswrapper[4904]: I0214 11:36:24.085370 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9243ea89-2a6d-43a8-85d7-b83b3e664680-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9243ea89-2a6d-43a8-85d7-b83b3e664680" (UID: "9243ea89-2a6d-43a8-85d7-b83b3e664680"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:36:24 crc kubenswrapper[4904]: I0214 11:36:24.142750 4904 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9243ea89-2a6d-43a8-85d7-b83b3e664680-inventory\") on node \"crc\" DevicePath \"\"" Feb 14 11:36:24 crc kubenswrapper[4904]: I0214 11:36:24.142788 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7w72p\" (UniqueName: \"kubernetes.io/projected/9243ea89-2a6d-43a8-85d7-b83b3e664680-kube-api-access-7w72p\") on node \"crc\" DevicePath \"\"" Feb 14 11:36:24 crc kubenswrapper[4904]: I0214 11:36:24.142802 4904 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9243ea89-2a6d-43a8-85d7-b83b3e664680-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 14 11:36:24 crc kubenswrapper[4904]: I0214 11:36:24.142810 4904 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9243ea89-2a6d-43a8-85d7-b83b3e664680-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:36:24 crc kubenswrapper[4904]: I0214 11:36:24.477709 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" event={"ID":"9243ea89-2a6d-43a8-85d7-b83b3e664680","Type":"ContainerDied","Data":"b38711e376ec0dc52976dec50dcce1552948b13bf91d1d5a5a275eed0734d85c"} Feb 14 11:36:24 crc kubenswrapper[4904]: I0214 11:36:24.477751 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b38711e376ec0dc52976dec50dcce1552948b13bf91d1d5a5a275eed0734d85c" Feb 14 11:36:24 crc kubenswrapper[4904]: I0214 11:36:24.477956 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl" Feb 14 11:36:25 crc kubenswrapper[4904]: I0214 11:36:25.847806 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="042a40bf-ba7e-4f1a-afb9-a4f82a3470e1" path="/var/lib/kubelet/pods/042a40bf-ba7e-4f1a-afb9-a4f82a3470e1/volumes" Feb 14 11:36:28 crc kubenswrapper[4904]: I0214 11:36:28.028140 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-jc69g"] Feb 14 11:36:28 crc kubenswrapper[4904]: I0214 11:36:28.043151 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-jc69g"] Feb 14 11:36:29 crc kubenswrapper[4904]: I0214 11:36:29.847250 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9a1a642-2b0a-4a5c-825f-883de77215e2" path="/var/lib/kubelet/pods/e9a1a642-2b0a-4a5c-825f-883de77215e2/volumes" Feb 14 11:36:32 crc kubenswrapper[4904]: I0214 11:36:32.836488 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:36:32 crc kubenswrapper[4904]: E0214 11:36:32.837877 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.025649 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj"] Feb 14 11:36:42 crc kubenswrapper[4904]: E0214 11:36:42.026383 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9243ea89-2a6d-43a8-85d7-b83b3e664680" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.026398 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="9243ea89-2a6d-43a8-85d7-b83b3e664680" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.026583 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="9243ea89-2a6d-43a8-85d7-b83b3e664680" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.027272 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.030222 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.030392 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-48rnz" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.030590 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.040357 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.046310 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj"] Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.093496 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4gpv\" (UniqueName: \"kubernetes.io/projected/80fcf5ac-b576-4551-8478-d411430ad030-kube-api-access-q4gpv\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj\" (UID: \"80fcf5ac-b576-4551-8478-d411430ad030\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.093630 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/80fcf5ac-b576-4551-8478-d411430ad030-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj\" (UID: \"80fcf5ac-b576-4551-8478-d411430ad030\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.093660 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80fcf5ac-b576-4551-8478-d411430ad030-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj\" (UID: \"80fcf5ac-b576-4551-8478-d411430ad030\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.093713 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80fcf5ac-b576-4551-8478-d411430ad030-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj\" (UID: \"80fcf5ac-b576-4551-8478-d411430ad030\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.195356 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/80fcf5ac-b576-4551-8478-d411430ad030-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj\" (UID: \"80fcf5ac-b576-4551-8478-d411430ad030\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.195425 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80fcf5ac-b576-4551-8478-d411430ad030-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj\" (UID: \"80fcf5ac-b576-4551-8478-d411430ad030\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.195495 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80fcf5ac-b576-4551-8478-d411430ad030-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj\" (UID: \"80fcf5ac-b576-4551-8478-d411430ad030\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.195572 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4gpv\" (UniqueName: \"kubernetes.io/projected/80fcf5ac-b576-4551-8478-d411430ad030-kube-api-access-q4gpv\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj\" (UID: \"80fcf5ac-b576-4551-8478-d411430ad030\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.202477 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/80fcf5ac-b576-4551-8478-d411430ad030-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj\" (UID: \"80fcf5ac-b576-4551-8478-d411430ad030\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.203358 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80fcf5ac-b576-4551-8478-d411430ad030-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj\" (UID: \"80fcf5ac-b576-4551-8478-d411430ad030\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.209175 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80fcf5ac-b576-4551-8478-d411430ad030-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj\" (UID: \"80fcf5ac-b576-4551-8478-d411430ad030\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.211322 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4gpv\" (UniqueName: \"kubernetes.io/projected/80fcf5ac-b576-4551-8478-d411430ad030-kube-api-access-q4gpv\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj\" (UID: \"80fcf5ac-b576-4551-8478-d411430ad030\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.345259 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.857432 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj"] Feb 14 11:36:42 crc kubenswrapper[4904]: I0214 11:36:42.861323 4904 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 14 11:36:43 crc kubenswrapper[4904]: I0214 11:36:43.644183 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" event={"ID":"80fcf5ac-b576-4551-8478-d411430ad030","Type":"ContainerStarted","Data":"747c70f7a52b504c3d376ef4f28838450a38749ea1c4b9be66b060f37274963c"} Feb 14 11:36:43 crc kubenswrapper[4904]: I0214 11:36:43.644528 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" event={"ID":"80fcf5ac-b576-4551-8478-d411430ad030","Type":"ContainerStarted","Data":"fa09f6f67314090aa6b0e6ee4033fe52073cb231299066e8811651b4ecd2de89"} Feb 14 11:36:43 crc kubenswrapper[4904]: I0214 11:36:43.667984 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" podStartSLOduration=1.264565908 podStartE2EDuration="1.667959131s" podCreationTimestamp="2026-02-14 11:36:42 +0000 UTC" firstStartedPulling="2026-02-14 11:36:42.861021531 +0000 UTC m=+1593.673786192" lastFinishedPulling="2026-02-14 11:36:43.264414754 +0000 UTC m=+1594.077179415" observedRunningTime="2026-02-14 11:36:43.658147583 +0000 UTC m=+1594.470912244" watchObservedRunningTime="2026-02-14 11:36:43.667959131 +0000 UTC m=+1594.480723812" Feb 14 11:36:44 crc kubenswrapper[4904]: I0214 11:36:44.837926 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:36:44 crc kubenswrapper[4904]: E0214 11:36:44.839555 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:36:47 crc kubenswrapper[4904]: I0214 11:36:47.658038 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mgzbf"] Feb 14 11:36:47 crc kubenswrapper[4904]: I0214 11:36:47.664438 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mgzbf" Feb 14 11:36:47 crc kubenswrapper[4904]: I0214 11:36:47.680753 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mgzbf"] Feb 14 11:36:47 crc kubenswrapper[4904]: I0214 11:36:47.707576 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92dc2e95-c354-4388-9e09-2470bfce8e34-utilities\") pod \"redhat-operators-mgzbf\" (UID: \"92dc2e95-c354-4388-9e09-2470bfce8e34\") " pod="openshift-marketplace/redhat-operators-mgzbf" Feb 14 11:36:47 crc kubenswrapper[4904]: I0214 11:36:47.707628 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92dc2e95-c354-4388-9e09-2470bfce8e34-catalog-content\") pod \"redhat-operators-mgzbf\" (UID: \"92dc2e95-c354-4388-9e09-2470bfce8e34\") " pod="openshift-marketplace/redhat-operators-mgzbf" Feb 14 11:36:47 crc kubenswrapper[4904]: I0214 11:36:47.707697 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2jxz\" (UniqueName: \"kubernetes.io/projected/92dc2e95-c354-4388-9e09-2470bfce8e34-kube-api-access-s2jxz\") pod \"redhat-operators-mgzbf\" (UID: \"92dc2e95-c354-4388-9e09-2470bfce8e34\") " pod="openshift-marketplace/redhat-operators-mgzbf" Feb 14 11:36:47 crc kubenswrapper[4904]: I0214 11:36:47.808817 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92dc2e95-c354-4388-9e09-2470bfce8e34-utilities\") pod \"redhat-operators-mgzbf\" (UID: \"92dc2e95-c354-4388-9e09-2470bfce8e34\") " pod="openshift-marketplace/redhat-operators-mgzbf" Feb 14 11:36:47 crc kubenswrapper[4904]: I0214 11:36:47.808960 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92dc2e95-c354-4388-9e09-2470bfce8e34-catalog-content\") pod \"redhat-operators-mgzbf\" (UID: \"92dc2e95-c354-4388-9e09-2470bfce8e34\") " pod="openshift-marketplace/redhat-operators-mgzbf" Feb 14 11:36:47 crc kubenswrapper[4904]: I0214 11:36:47.809038 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2jxz\" (UniqueName: \"kubernetes.io/projected/92dc2e95-c354-4388-9e09-2470bfce8e34-kube-api-access-s2jxz\") pod \"redhat-operators-mgzbf\" (UID: \"92dc2e95-c354-4388-9e09-2470bfce8e34\") " pod="openshift-marketplace/redhat-operators-mgzbf" Feb 14 11:36:47 crc kubenswrapper[4904]: I0214 11:36:47.809277 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92dc2e95-c354-4388-9e09-2470bfce8e34-utilities\") pod \"redhat-operators-mgzbf\" (UID: \"92dc2e95-c354-4388-9e09-2470bfce8e34\") " pod="openshift-marketplace/redhat-operators-mgzbf" Feb 14 11:36:47 crc kubenswrapper[4904]: I0214 11:36:47.809642 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92dc2e95-c354-4388-9e09-2470bfce8e34-catalog-content\") pod \"redhat-operators-mgzbf\" (UID: \"92dc2e95-c354-4388-9e09-2470bfce8e34\") " pod="openshift-marketplace/redhat-operators-mgzbf" Feb 14 11:36:47 crc kubenswrapper[4904]: I0214 11:36:47.846720 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2jxz\" (UniqueName: \"kubernetes.io/projected/92dc2e95-c354-4388-9e09-2470bfce8e34-kube-api-access-s2jxz\") pod \"redhat-operators-mgzbf\" (UID: \"92dc2e95-c354-4388-9e09-2470bfce8e34\") " pod="openshift-marketplace/redhat-operators-mgzbf" Feb 14 11:36:48 crc kubenswrapper[4904]: I0214 11:36:48.003147 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mgzbf" Feb 14 11:36:48 crc kubenswrapper[4904]: I0214 11:36:48.484309 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mgzbf"] Feb 14 11:36:48 crc kubenswrapper[4904]: I0214 11:36:48.685877 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgzbf" event={"ID":"92dc2e95-c354-4388-9e09-2470bfce8e34","Type":"ContainerStarted","Data":"8d132abc2863da2908ec213a74b0f898f8908a054768a6bfb9cab1a0de611142"} Feb 14 11:36:48 crc kubenswrapper[4904]: I0214 11:36:48.685918 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgzbf" event={"ID":"92dc2e95-c354-4388-9e09-2470bfce8e34","Type":"ContainerStarted","Data":"aa1b4d185f5062b2e2eca8211a38cc3a939354a164370b0956cb2a7e6a2de0e2"} Feb 14 11:36:49 crc kubenswrapper[4904]: I0214 11:36:49.694674 4904 generic.go:334] "Generic (PLEG): container finished" podID="92dc2e95-c354-4388-9e09-2470bfce8e34" containerID="8d132abc2863da2908ec213a74b0f898f8908a054768a6bfb9cab1a0de611142" exitCode=0 Feb 14 11:36:49 crc kubenswrapper[4904]: I0214 11:36:49.694774 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgzbf" event={"ID":"92dc2e95-c354-4388-9e09-2470bfce8e34","Type":"ContainerDied","Data":"8d132abc2863da2908ec213a74b0f898f8908a054768a6bfb9cab1a0de611142"} Feb 14 11:36:50 crc kubenswrapper[4904]: I0214 11:36:50.706482 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgzbf" event={"ID":"92dc2e95-c354-4388-9e09-2470bfce8e34","Type":"ContainerStarted","Data":"95f04c07ecf054c17691367595f1da2dc82da0cc508291e2ebd6762b40386d2d"} Feb 14 11:36:52 crc kubenswrapper[4904]: I0214 11:36:52.433816 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p7jkw"] Feb 14 11:36:52 crc kubenswrapper[4904]: I0214 11:36:52.435854 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p7jkw" Feb 14 11:36:52 crc kubenswrapper[4904]: I0214 11:36:52.449592 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p7jkw"] Feb 14 11:36:52 crc kubenswrapper[4904]: I0214 11:36:52.496181 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0aea853-283c-4da1-9dfb-f47bd4ed1117-utilities\") pod \"certified-operators-p7jkw\" (UID: \"b0aea853-283c-4da1-9dfb-f47bd4ed1117\") " pod="openshift-marketplace/certified-operators-p7jkw" Feb 14 11:36:52 crc kubenswrapper[4904]: I0214 11:36:52.496250 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc922\" (UniqueName: \"kubernetes.io/projected/b0aea853-283c-4da1-9dfb-f47bd4ed1117-kube-api-access-mc922\") pod \"certified-operators-p7jkw\" (UID: \"b0aea853-283c-4da1-9dfb-f47bd4ed1117\") " pod="openshift-marketplace/certified-operators-p7jkw" Feb 14 11:36:52 crc kubenswrapper[4904]: I0214 11:36:52.496331 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0aea853-283c-4da1-9dfb-f47bd4ed1117-catalog-content\") pod \"certified-operators-p7jkw\" (UID: \"b0aea853-283c-4da1-9dfb-f47bd4ed1117\") " pod="openshift-marketplace/certified-operators-p7jkw" Feb 14 11:36:52 crc kubenswrapper[4904]: I0214 11:36:52.597960 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0aea853-283c-4da1-9dfb-f47bd4ed1117-catalog-content\") pod \"certified-operators-p7jkw\" (UID: \"b0aea853-283c-4da1-9dfb-f47bd4ed1117\") " pod="openshift-marketplace/certified-operators-p7jkw" Feb 14 11:36:52 crc kubenswrapper[4904]: I0214 11:36:52.598055 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0aea853-283c-4da1-9dfb-f47bd4ed1117-utilities\") pod \"certified-operators-p7jkw\" (UID: \"b0aea853-283c-4da1-9dfb-f47bd4ed1117\") " pod="openshift-marketplace/certified-operators-p7jkw" Feb 14 11:36:52 crc kubenswrapper[4904]: I0214 11:36:52.598129 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc922\" (UniqueName: \"kubernetes.io/projected/b0aea853-283c-4da1-9dfb-f47bd4ed1117-kube-api-access-mc922\") pod \"certified-operators-p7jkw\" (UID: \"b0aea853-283c-4da1-9dfb-f47bd4ed1117\") " pod="openshift-marketplace/certified-operators-p7jkw" Feb 14 11:36:52 crc kubenswrapper[4904]: I0214 11:36:52.599038 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0aea853-283c-4da1-9dfb-f47bd4ed1117-utilities\") pod \"certified-operators-p7jkw\" (UID: \"b0aea853-283c-4da1-9dfb-f47bd4ed1117\") " pod="openshift-marketplace/certified-operators-p7jkw" Feb 14 11:36:52 crc kubenswrapper[4904]: I0214 11:36:52.599215 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0aea853-283c-4da1-9dfb-f47bd4ed1117-catalog-content\") pod \"certified-operators-p7jkw\" (UID: \"b0aea853-283c-4da1-9dfb-f47bd4ed1117\") " pod="openshift-marketplace/certified-operators-p7jkw" Feb 14 11:36:52 crc kubenswrapper[4904]: I0214 11:36:52.647489 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc922\" (UniqueName: \"kubernetes.io/projected/b0aea853-283c-4da1-9dfb-f47bd4ed1117-kube-api-access-mc922\") pod \"certified-operators-p7jkw\" (UID: \"b0aea853-283c-4da1-9dfb-f47bd4ed1117\") " pod="openshift-marketplace/certified-operators-p7jkw" Feb 14 11:36:52 crc kubenswrapper[4904]: I0214 11:36:52.769136 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p7jkw" Feb 14 11:36:53 crc kubenswrapper[4904]: I0214 11:36:53.710601 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p7jkw"] Feb 14 11:36:53 crc kubenswrapper[4904]: I0214 11:36:53.740563 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7jkw" event={"ID":"b0aea853-283c-4da1-9dfb-f47bd4ed1117","Type":"ContainerStarted","Data":"23ccab634ffe52f8ec943e0e3044f60ec0bc74219c6d7340bbbe6697c32da36f"} Feb 14 11:36:54 crc kubenswrapper[4904]: I0214 11:36:54.749883 4904 generic.go:334] "Generic (PLEG): container finished" podID="b0aea853-283c-4da1-9dfb-f47bd4ed1117" containerID="a3717cfc797fa302fe433c2d44fe99f2d798bbd9250402c1313591a7095a92d1" exitCode=0 Feb 14 11:36:54 crc kubenswrapper[4904]: I0214 11:36:54.749947 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7jkw" event={"ID":"b0aea853-283c-4da1-9dfb-f47bd4ed1117","Type":"ContainerDied","Data":"a3717cfc797fa302fe433c2d44fe99f2d798bbd9250402c1313591a7095a92d1"} Feb 14 11:36:56 crc kubenswrapper[4904]: I0214 11:36:56.778524 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7jkw" event={"ID":"b0aea853-283c-4da1-9dfb-f47bd4ed1117","Type":"ContainerStarted","Data":"a6f4e650e993b09d2866943bf5e35846a9c8d6219f3cf53ae930b36c258c73e9"} Feb 14 11:36:57 crc kubenswrapper[4904]: I0214 11:36:57.788880 4904 generic.go:334] "Generic (PLEG): container finished" podID="92dc2e95-c354-4388-9e09-2470bfce8e34" containerID="95f04c07ecf054c17691367595f1da2dc82da0cc508291e2ebd6762b40386d2d" exitCode=0 Feb 14 11:36:57 crc kubenswrapper[4904]: I0214 11:36:57.788976 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgzbf" event={"ID":"92dc2e95-c354-4388-9e09-2470bfce8e34","Type":"ContainerDied","Data":"95f04c07ecf054c17691367595f1da2dc82da0cc508291e2ebd6762b40386d2d"} Feb 14 11:36:58 crc kubenswrapper[4904]: I0214 11:36:58.798074 4904 generic.go:334] "Generic (PLEG): container finished" podID="b0aea853-283c-4da1-9dfb-f47bd4ed1117" containerID="a6f4e650e993b09d2866943bf5e35846a9c8d6219f3cf53ae930b36c258c73e9" exitCode=0 Feb 14 11:36:58 crc kubenswrapper[4904]: I0214 11:36:58.798151 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7jkw" event={"ID":"b0aea853-283c-4da1-9dfb-f47bd4ed1117","Type":"ContainerDied","Data":"a6f4e650e993b09d2866943bf5e35846a9c8d6219f3cf53ae930b36c258c73e9"} Feb 14 11:36:58 crc kubenswrapper[4904]: I0214 11:36:58.802678 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgzbf" event={"ID":"92dc2e95-c354-4388-9e09-2470bfce8e34","Type":"ContainerStarted","Data":"5a81569b8f03b6597798d91683cc04e77566b07fd769aeb609afd5ce9805e44f"} Feb 14 11:36:58 crc kubenswrapper[4904]: I0214 11:36:58.837711 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:36:58 crc kubenswrapper[4904]: E0214 11:36:58.838158 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:36:58 crc kubenswrapper[4904]: I0214 11:36:58.850699 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mgzbf" podStartSLOduration=3.276913084 podStartE2EDuration="11.850680702s" podCreationTimestamp="2026-02-14 11:36:47 +0000 UTC" firstStartedPulling="2026-02-14 11:36:49.696711905 +0000 UTC m=+1600.509476566" lastFinishedPulling="2026-02-14 11:36:58.270479523 +0000 UTC m=+1609.083244184" observedRunningTime="2026-02-14 11:36:58.835581458 +0000 UTC m=+1609.648346119" watchObservedRunningTime="2026-02-14 11:36:58.850680702 +0000 UTC m=+1609.663445363" Feb 14 11:36:59 crc kubenswrapper[4904]: I0214 11:36:59.812755 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7jkw" event={"ID":"b0aea853-283c-4da1-9dfb-f47bd4ed1117","Type":"ContainerStarted","Data":"b568f476e4f74ad62d661e7563b2754fe291347f5098d94e8653914cf5093858"} Feb 14 11:36:59 crc kubenswrapper[4904]: I0214 11:36:59.832198 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p7jkw" podStartSLOduration=3.382307984 podStartE2EDuration="7.832181506s" podCreationTimestamp="2026-02-14 11:36:52 +0000 UTC" firstStartedPulling="2026-02-14 11:36:54.752006043 +0000 UTC m=+1605.564770694" lastFinishedPulling="2026-02-14 11:36:59.201879555 +0000 UTC m=+1610.014644216" observedRunningTime="2026-02-14 11:36:59.831339302 +0000 UTC m=+1610.644103963" watchObservedRunningTime="2026-02-14 11:36:59.832181506 +0000 UTC m=+1610.644946167" Feb 14 11:37:02 crc kubenswrapper[4904]: I0214 11:37:02.769999 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p7jkw" Feb 14 11:37:02 crc kubenswrapper[4904]: I0214 11:37:02.772604 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p7jkw" Feb 14 11:37:03 crc kubenswrapper[4904]: I0214 11:37:03.813774 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-p7jkw" podUID="b0aea853-283c-4da1-9dfb-f47bd4ed1117" containerName="registry-server" probeResult="failure" output=< Feb 14 11:37:03 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 11:37:03 crc kubenswrapper[4904]: > Feb 14 11:37:05 crc kubenswrapper[4904]: I0214 11:37:05.055882 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-t6bn6"] Feb 14 11:37:05 crc kubenswrapper[4904]: I0214 11:37:05.074261 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-t6bn6"] Feb 14 11:37:05 crc kubenswrapper[4904]: I0214 11:37:05.846088 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2abcf43f-6f3b-4804-83a5-ba42fc66f2ec" path="/var/lib/kubelet/pods/2abcf43f-6f3b-4804-83a5-ba42fc66f2ec/volumes" Feb 14 11:37:08 crc kubenswrapper[4904]: I0214 11:37:08.004016 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mgzbf" Feb 14 11:37:08 crc kubenswrapper[4904]: I0214 11:37:08.004353 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mgzbf" Feb 14 11:37:09 crc kubenswrapper[4904]: I0214 11:37:09.064499 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mgzbf" podUID="92dc2e95-c354-4388-9e09-2470bfce8e34" containerName="registry-server" probeResult="failure" output=< Feb 14 11:37:09 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 11:37:09 crc kubenswrapper[4904]: > Feb 14 11:37:10 crc kubenswrapper[4904]: I0214 11:37:10.839007 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:37:10 crc kubenswrapper[4904]: E0214 11:37:10.839554 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:37:12 crc kubenswrapper[4904]: I0214 11:37:12.836956 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p7jkw" Feb 14 11:37:12 crc kubenswrapper[4904]: I0214 11:37:12.888302 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p7jkw" Feb 14 11:37:13 crc kubenswrapper[4904]: I0214 11:37:13.083290 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p7jkw"] Feb 14 11:37:13 crc kubenswrapper[4904]: I0214 11:37:13.931260 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p7jkw" podUID="b0aea853-283c-4da1-9dfb-f47bd4ed1117" containerName="registry-server" containerID="cri-o://b568f476e4f74ad62d661e7563b2754fe291347f5098d94e8653914cf5093858" gracePeriod=2 Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.411630 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p7jkw" Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.582174 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0aea853-283c-4da1-9dfb-f47bd4ed1117-catalog-content\") pod \"b0aea853-283c-4da1-9dfb-f47bd4ed1117\" (UID: \"b0aea853-283c-4da1-9dfb-f47bd4ed1117\") " Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.582282 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mc922\" (UniqueName: \"kubernetes.io/projected/b0aea853-283c-4da1-9dfb-f47bd4ed1117-kube-api-access-mc922\") pod \"b0aea853-283c-4da1-9dfb-f47bd4ed1117\" (UID: \"b0aea853-283c-4da1-9dfb-f47bd4ed1117\") " Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.582397 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0aea853-283c-4da1-9dfb-f47bd4ed1117-utilities\") pod \"b0aea853-283c-4da1-9dfb-f47bd4ed1117\" (UID: \"b0aea853-283c-4da1-9dfb-f47bd4ed1117\") " Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.583208 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0aea853-283c-4da1-9dfb-f47bd4ed1117-utilities" (OuterVolumeSpecName: "utilities") pod "b0aea853-283c-4da1-9dfb-f47bd4ed1117" (UID: "b0aea853-283c-4da1-9dfb-f47bd4ed1117"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.589689 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0aea853-283c-4da1-9dfb-f47bd4ed1117-kube-api-access-mc922" (OuterVolumeSpecName: "kube-api-access-mc922") pod "b0aea853-283c-4da1-9dfb-f47bd4ed1117" (UID: "b0aea853-283c-4da1-9dfb-f47bd4ed1117"). InnerVolumeSpecName "kube-api-access-mc922". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.619863 4904 scope.go:117] "RemoveContainer" containerID="0e7c0155f80a4151ce218be32053768a5e7f953506fde9f13203f41b15c3c107" Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.634079 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0aea853-283c-4da1-9dfb-f47bd4ed1117-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0aea853-283c-4da1-9dfb-f47bd4ed1117" (UID: "b0aea853-283c-4da1-9dfb-f47bd4ed1117"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.657295 4904 scope.go:117] "RemoveContainer" containerID="41db167ff7dc3e24f9b2ac1e32c2405e16f5f00ee2611e44e5400ce8318cbd67" Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.684465 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0aea853-283c-4da1-9dfb-f47bd4ed1117-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.684500 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mc922\" (UniqueName: \"kubernetes.io/projected/b0aea853-283c-4da1-9dfb-f47bd4ed1117-kube-api-access-mc922\") on node \"crc\" DevicePath \"\"" Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.684512 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0aea853-283c-4da1-9dfb-f47bd4ed1117-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.701464 4904 scope.go:117] "RemoveContainer" containerID="bd7ace9778e88cca437bf1ade30faea54060444a0a0b802788627437c3702387" Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.721876 4904 scope.go:117] "RemoveContainer" containerID="ac18286729f4904e4ad9a3194b4bf19ab0d8831192515607e99661486cdffff4" Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.739044 4904 scope.go:117] "RemoveContainer" containerID="063736057a86ef2533b8d53d185c3959a9c8feed0f08ea11c82cc5ffeeb6c0e0" Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.940936 4904 generic.go:334] "Generic (PLEG): container finished" podID="b0aea853-283c-4da1-9dfb-f47bd4ed1117" containerID="b568f476e4f74ad62d661e7563b2754fe291347f5098d94e8653914cf5093858" exitCode=0 Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.940973 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7jkw" event={"ID":"b0aea853-283c-4da1-9dfb-f47bd4ed1117","Type":"ContainerDied","Data":"b568f476e4f74ad62d661e7563b2754fe291347f5098d94e8653914cf5093858"} Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.940981 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p7jkw" Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.940995 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7jkw" event={"ID":"b0aea853-283c-4da1-9dfb-f47bd4ed1117","Type":"ContainerDied","Data":"23ccab634ffe52f8ec943e0e3044f60ec0bc74219c6d7340bbbe6697c32da36f"} Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.941016 4904 scope.go:117] "RemoveContainer" containerID="b568f476e4f74ad62d661e7563b2754fe291347f5098d94e8653914cf5093858" Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.970195 4904 scope.go:117] "RemoveContainer" containerID="a6f4e650e993b09d2866943bf5e35846a9c8d6219f3cf53ae930b36c258c73e9" Feb 14 11:37:14 crc kubenswrapper[4904]: I0214 11:37:14.991622 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p7jkw"] Feb 14 11:37:15 crc kubenswrapper[4904]: I0214 11:37:15.007568 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p7jkw"] Feb 14 11:37:15 crc kubenswrapper[4904]: I0214 11:37:15.021762 4904 scope.go:117] "RemoveContainer" containerID="a3717cfc797fa302fe433c2d44fe99f2d798bbd9250402c1313591a7095a92d1" Feb 14 11:37:15 crc kubenswrapper[4904]: I0214 11:37:15.046752 4904 scope.go:117] "RemoveContainer" containerID="b568f476e4f74ad62d661e7563b2754fe291347f5098d94e8653914cf5093858" Feb 14 11:37:15 crc kubenswrapper[4904]: E0214 11:37:15.047142 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b568f476e4f74ad62d661e7563b2754fe291347f5098d94e8653914cf5093858\": container with ID starting with b568f476e4f74ad62d661e7563b2754fe291347f5098d94e8653914cf5093858 not found: ID does not exist" containerID="b568f476e4f74ad62d661e7563b2754fe291347f5098d94e8653914cf5093858" Feb 14 11:37:15 crc kubenswrapper[4904]: I0214 11:37:15.047179 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b568f476e4f74ad62d661e7563b2754fe291347f5098d94e8653914cf5093858"} err="failed to get container status \"b568f476e4f74ad62d661e7563b2754fe291347f5098d94e8653914cf5093858\": rpc error: code = NotFound desc = could not find container \"b568f476e4f74ad62d661e7563b2754fe291347f5098d94e8653914cf5093858\": container with ID starting with b568f476e4f74ad62d661e7563b2754fe291347f5098d94e8653914cf5093858 not found: ID does not exist" Feb 14 11:37:15 crc kubenswrapper[4904]: I0214 11:37:15.047208 4904 scope.go:117] "RemoveContainer" containerID="a6f4e650e993b09d2866943bf5e35846a9c8d6219f3cf53ae930b36c258c73e9" Feb 14 11:37:15 crc kubenswrapper[4904]: E0214 11:37:15.047445 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6f4e650e993b09d2866943bf5e35846a9c8d6219f3cf53ae930b36c258c73e9\": container with ID starting with a6f4e650e993b09d2866943bf5e35846a9c8d6219f3cf53ae930b36c258c73e9 not found: ID does not exist" containerID="a6f4e650e993b09d2866943bf5e35846a9c8d6219f3cf53ae930b36c258c73e9" Feb 14 11:37:15 crc kubenswrapper[4904]: I0214 11:37:15.047479 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6f4e650e993b09d2866943bf5e35846a9c8d6219f3cf53ae930b36c258c73e9"} err="failed to get container status \"a6f4e650e993b09d2866943bf5e35846a9c8d6219f3cf53ae930b36c258c73e9\": rpc error: code = NotFound desc = could not find container \"a6f4e650e993b09d2866943bf5e35846a9c8d6219f3cf53ae930b36c258c73e9\": container with ID starting with a6f4e650e993b09d2866943bf5e35846a9c8d6219f3cf53ae930b36c258c73e9 not found: ID does not exist" Feb 14 11:37:15 crc kubenswrapper[4904]: I0214 11:37:15.047539 4904 scope.go:117] "RemoveContainer" containerID="a3717cfc797fa302fe433c2d44fe99f2d798bbd9250402c1313591a7095a92d1" Feb 14 11:37:15 crc kubenswrapper[4904]: E0214 11:37:15.048155 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3717cfc797fa302fe433c2d44fe99f2d798bbd9250402c1313591a7095a92d1\": container with ID starting with a3717cfc797fa302fe433c2d44fe99f2d798bbd9250402c1313591a7095a92d1 not found: ID does not exist" containerID="a3717cfc797fa302fe433c2d44fe99f2d798bbd9250402c1313591a7095a92d1" Feb 14 11:37:15 crc kubenswrapper[4904]: I0214 11:37:15.048220 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3717cfc797fa302fe433c2d44fe99f2d798bbd9250402c1313591a7095a92d1"} err="failed to get container status \"a3717cfc797fa302fe433c2d44fe99f2d798bbd9250402c1313591a7095a92d1\": rpc error: code = NotFound desc = could not find container \"a3717cfc797fa302fe433c2d44fe99f2d798bbd9250402c1313591a7095a92d1\": container with ID starting with a3717cfc797fa302fe433c2d44fe99f2d798bbd9250402c1313591a7095a92d1 not found: ID does not exist" Feb 14 11:37:15 crc kubenswrapper[4904]: I0214 11:37:15.848058 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0aea853-283c-4da1-9dfb-f47bd4ed1117" path="/var/lib/kubelet/pods/b0aea853-283c-4da1-9dfb-f47bd4ed1117/volumes" Feb 14 11:37:18 crc kubenswrapper[4904]: I0214 11:37:18.039226 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-v6zmt"] Feb 14 11:37:18 crc kubenswrapper[4904]: I0214 11:37:18.046719 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-ft76h"] Feb 14 11:37:18 crc kubenswrapper[4904]: I0214 11:37:18.054498 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-ft76h"] Feb 14 11:37:18 crc kubenswrapper[4904]: I0214 11:37:18.062760 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-v6zmt"] Feb 14 11:37:18 crc kubenswrapper[4904]: I0214 11:37:18.081820 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mgzbf" Feb 14 11:37:18 crc kubenswrapper[4904]: I0214 11:37:18.144698 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mgzbf" Feb 14 11:37:18 crc kubenswrapper[4904]: I0214 11:37:18.489087 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mgzbf"] Feb 14 11:37:19 crc kubenswrapper[4904]: I0214 11:37:19.850782 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60ac5de4-7c8e-4065-a846-3de7a3cffcc0" path="/var/lib/kubelet/pods/60ac5de4-7c8e-4065-a846-3de7a3cffcc0/volumes" Feb 14 11:37:19 crc kubenswrapper[4904]: I0214 11:37:19.854383 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec37844d-1c49-4aa6-b9e1-87ba60eadbe1" path="/var/lib/kubelet/pods/ec37844d-1c49-4aa6-b9e1-87ba60eadbe1/volumes" Feb 14 11:37:19 crc kubenswrapper[4904]: I0214 11:37:19.997143 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mgzbf" podUID="92dc2e95-c354-4388-9e09-2470bfce8e34" containerName="registry-server" containerID="cri-o://5a81569b8f03b6597798d91683cc04e77566b07fd769aeb609afd5ce9805e44f" gracePeriod=2 Feb 14 11:37:20 crc kubenswrapper[4904]: I0214 11:37:20.468138 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mgzbf" Feb 14 11:37:20 crc kubenswrapper[4904]: I0214 11:37:20.600088 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92dc2e95-c354-4388-9e09-2470bfce8e34-catalog-content\") pod \"92dc2e95-c354-4388-9e09-2470bfce8e34\" (UID: \"92dc2e95-c354-4388-9e09-2470bfce8e34\") " Feb 14 11:37:20 crc kubenswrapper[4904]: I0214 11:37:20.600372 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92dc2e95-c354-4388-9e09-2470bfce8e34-utilities\") pod \"92dc2e95-c354-4388-9e09-2470bfce8e34\" (UID: \"92dc2e95-c354-4388-9e09-2470bfce8e34\") " Feb 14 11:37:20 crc kubenswrapper[4904]: I0214 11:37:20.600420 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2jxz\" (UniqueName: \"kubernetes.io/projected/92dc2e95-c354-4388-9e09-2470bfce8e34-kube-api-access-s2jxz\") pod \"92dc2e95-c354-4388-9e09-2470bfce8e34\" (UID: \"92dc2e95-c354-4388-9e09-2470bfce8e34\") " Feb 14 11:37:20 crc kubenswrapper[4904]: I0214 11:37:20.601287 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92dc2e95-c354-4388-9e09-2470bfce8e34-utilities" (OuterVolumeSpecName: "utilities") pod "92dc2e95-c354-4388-9e09-2470bfce8e34" (UID: "92dc2e95-c354-4388-9e09-2470bfce8e34"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:37:20 crc kubenswrapper[4904]: I0214 11:37:20.610823 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92dc2e95-c354-4388-9e09-2470bfce8e34-kube-api-access-s2jxz" (OuterVolumeSpecName: "kube-api-access-s2jxz") pod "92dc2e95-c354-4388-9e09-2470bfce8e34" (UID: "92dc2e95-c354-4388-9e09-2470bfce8e34"). InnerVolumeSpecName "kube-api-access-s2jxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:37:20 crc kubenswrapper[4904]: I0214 11:37:20.705939 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92dc2e95-c354-4388-9e09-2470bfce8e34-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:37:20 crc kubenswrapper[4904]: I0214 11:37:20.706548 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2jxz\" (UniqueName: \"kubernetes.io/projected/92dc2e95-c354-4388-9e09-2470bfce8e34-kube-api-access-s2jxz\") on node \"crc\" DevicePath \"\"" Feb 14 11:37:20 crc kubenswrapper[4904]: I0214 11:37:20.740664 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92dc2e95-c354-4388-9e09-2470bfce8e34-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92dc2e95-c354-4388-9e09-2470bfce8e34" (UID: "92dc2e95-c354-4388-9e09-2470bfce8e34"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:37:20 crc kubenswrapper[4904]: I0214 11:37:20.808453 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92dc2e95-c354-4388-9e09-2470bfce8e34-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:37:21 crc kubenswrapper[4904]: I0214 11:37:21.007548 4904 generic.go:334] "Generic (PLEG): container finished" podID="92dc2e95-c354-4388-9e09-2470bfce8e34" containerID="5a81569b8f03b6597798d91683cc04e77566b07fd769aeb609afd5ce9805e44f" exitCode=0 Feb 14 11:37:21 crc kubenswrapper[4904]: I0214 11:37:21.007597 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgzbf" event={"ID":"92dc2e95-c354-4388-9e09-2470bfce8e34","Type":"ContainerDied","Data":"5a81569b8f03b6597798d91683cc04e77566b07fd769aeb609afd5ce9805e44f"} Feb 14 11:37:21 crc kubenswrapper[4904]: I0214 11:37:21.007627 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgzbf" event={"ID":"92dc2e95-c354-4388-9e09-2470bfce8e34","Type":"ContainerDied","Data":"aa1b4d185f5062b2e2eca8211a38cc3a939354a164370b0956cb2a7e6a2de0e2"} Feb 14 11:37:21 crc kubenswrapper[4904]: I0214 11:37:21.007643 4904 scope.go:117] "RemoveContainer" containerID="5a81569b8f03b6597798d91683cc04e77566b07fd769aeb609afd5ce9805e44f" Feb 14 11:37:21 crc kubenswrapper[4904]: I0214 11:37:21.007678 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mgzbf" Feb 14 11:37:21 crc kubenswrapper[4904]: I0214 11:37:21.029051 4904 scope.go:117] "RemoveContainer" containerID="95f04c07ecf054c17691367595f1da2dc82da0cc508291e2ebd6762b40386d2d" Feb 14 11:37:21 crc kubenswrapper[4904]: I0214 11:37:21.068133 4904 scope.go:117] "RemoveContainer" containerID="8d132abc2863da2908ec213a74b0f898f8908a054768a6bfb9cab1a0de611142" Feb 14 11:37:21 crc kubenswrapper[4904]: I0214 11:37:21.068482 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mgzbf"] Feb 14 11:37:21 crc kubenswrapper[4904]: I0214 11:37:21.077718 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mgzbf"] Feb 14 11:37:21 crc kubenswrapper[4904]: I0214 11:37:21.108750 4904 scope.go:117] "RemoveContainer" containerID="5a81569b8f03b6597798d91683cc04e77566b07fd769aeb609afd5ce9805e44f" Feb 14 11:37:21 crc kubenswrapper[4904]: E0214 11:37:21.109315 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a81569b8f03b6597798d91683cc04e77566b07fd769aeb609afd5ce9805e44f\": container with ID starting with 5a81569b8f03b6597798d91683cc04e77566b07fd769aeb609afd5ce9805e44f not found: ID does not exist" containerID="5a81569b8f03b6597798d91683cc04e77566b07fd769aeb609afd5ce9805e44f" Feb 14 11:37:21 crc kubenswrapper[4904]: I0214 11:37:21.109406 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a81569b8f03b6597798d91683cc04e77566b07fd769aeb609afd5ce9805e44f"} err="failed to get container status \"5a81569b8f03b6597798d91683cc04e77566b07fd769aeb609afd5ce9805e44f\": rpc error: code = NotFound desc = could not find container \"5a81569b8f03b6597798d91683cc04e77566b07fd769aeb609afd5ce9805e44f\": container with ID starting with 5a81569b8f03b6597798d91683cc04e77566b07fd769aeb609afd5ce9805e44f not found: ID does not exist" Feb 14 11:37:21 crc kubenswrapper[4904]: I0214 11:37:21.109457 4904 scope.go:117] "RemoveContainer" containerID="95f04c07ecf054c17691367595f1da2dc82da0cc508291e2ebd6762b40386d2d" Feb 14 11:37:21 crc kubenswrapper[4904]: E0214 11:37:21.110050 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95f04c07ecf054c17691367595f1da2dc82da0cc508291e2ebd6762b40386d2d\": container with ID starting with 95f04c07ecf054c17691367595f1da2dc82da0cc508291e2ebd6762b40386d2d not found: ID does not exist" containerID="95f04c07ecf054c17691367595f1da2dc82da0cc508291e2ebd6762b40386d2d" Feb 14 11:37:21 crc kubenswrapper[4904]: I0214 11:37:21.110078 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95f04c07ecf054c17691367595f1da2dc82da0cc508291e2ebd6762b40386d2d"} err="failed to get container status \"95f04c07ecf054c17691367595f1da2dc82da0cc508291e2ebd6762b40386d2d\": rpc error: code = NotFound desc = could not find container \"95f04c07ecf054c17691367595f1da2dc82da0cc508291e2ebd6762b40386d2d\": container with ID starting with 95f04c07ecf054c17691367595f1da2dc82da0cc508291e2ebd6762b40386d2d not found: ID does not exist" Feb 14 11:37:21 crc kubenswrapper[4904]: I0214 11:37:21.110100 4904 scope.go:117] "RemoveContainer" containerID="8d132abc2863da2908ec213a74b0f898f8908a054768a6bfb9cab1a0de611142" Feb 14 11:37:21 crc kubenswrapper[4904]: E0214 11:37:21.110472 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d132abc2863da2908ec213a74b0f898f8908a054768a6bfb9cab1a0de611142\": container with ID starting with 8d132abc2863da2908ec213a74b0f898f8908a054768a6bfb9cab1a0de611142 not found: ID does not exist" containerID="8d132abc2863da2908ec213a74b0f898f8908a054768a6bfb9cab1a0de611142" Feb 14 11:37:21 crc kubenswrapper[4904]: I0214 11:37:21.110505 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d132abc2863da2908ec213a74b0f898f8908a054768a6bfb9cab1a0de611142"} err="failed to get container status \"8d132abc2863da2908ec213a74b0f898f8908a054768a6bfb9cab1a0de611142\": rpc error: code = NotFound desc = could not find container \"8d132abc2863da2908ec213a74b0f898f8908a054768a6bfb9cab1a0de611142\": container with ID starting with 8d132abc2863da2908ec213a74b0f898f8908a054768a6bfb9cab1a0de611142 not found: ID does not exist" Feb 14 11:37:21 crc kubenswrapper[4904]: I0214 11:37:21.847984 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92dc2e95-c354-4388-9e09-2470bfce8e34" path="/var/lib/kubelet/pods/92dc2e95-c354-4388-9e09-2470bfce8e34/volumes" Feb 14 11:37:22 crc kubenswrapper[4904]: I0214 11:37:22.835941 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:37:22 crc kubenswrapper[4904]: E0214 11:37:22.836460 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:37:30 crc kubenswrapper[4904]: I0214 11:37:30.034553 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-nnlq9"] Feb 14 11:37:30 crc kubenswrapper[4904]: I0214 11:37:30.047324 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-nnlq9"] Feb 14 11:37:31 crc kubenswrapper[4904]: I0214 11:37:31.027087 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-sflms"] Feb 14 11:37:31 crc kubenswrapper[4904]: I0214 11:37:31.046898 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-sflms"] Feb 14 11:37:31 crc kubenswrapper[4904]: I0214 11:37:31.845810 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b349fb1-5e71-45ff-88e9-03592bfe2504" path="/var/lib/kubelet/pods/6b349fb1-5e71-45ff-88e9-03592bfe2504/volumes" Feb 14 11:37:31 crc kubenswrapper[4904]: I0214 11:37:31.847512 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c698a36a-d049-4fdb-b279-1b62d1f807d3" path="/var/lib/kubelet/pods/c698a36a-d049-4fdb-b279-1b62d1f807d3/volumes" Feb 14 11:37:35 crc kubenswrapper[4904]: I0214 11:37:35.837037 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:37:35 crc kubenswrapper[4904]: E0214 11:37:35.837781 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:37:49 crc kubenswrapper[4904]: I0214 11:37:49.837006 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:37:49 crc kubenswrapper[4904]: E0214 11:37:49.838105 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:38:03 crc kubenswrapper[4904]: I0214 11:38:03.836723 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:38:03 crc kubenswrapper[4904]: E0214 11:38:03.837478 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:38:14 crc kubenswrapper[4904]: I0214 11:38:14.837223 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:38:14 crc kubenswrapper[4904]: E0214 11:38:14.838117 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:38:14 crc kubenswrapper[4904]: I0214 11:38:14.840508 4904 scope.go:117] "RemoveContainer" containerID="752aa5a87f01259b48e8dea972209d56a5ec87f518b1e02f7097ac9ce281ea76" Feb 14 11:38:14 crc kubenswrapper[4904]: I0214 11:38:14.862432 4904 scope.go:117] "RemoveContainer" containerID="b37fbd340585a792ef4d238687d76139e067db8414b45090fb4d705a9b528d75" Feb 14 11:38:14 crc kubenswrapper[4904]: I0214 11:38:14.919702 4904 scope.go:117] "RemoveContainer" containerID="2f654fcc93b3f55b28a5fe2cb572abfbf0d7e75bf3484ef586fdda4fcdd6723b" Feb 14 11:38:14 crc kubenswrapper[4904]: I0214 11:38:14.949787 4904 scope.go:117] "RemoveContainer" containerID="4d14917974e44826233222d2c4beb085c33e448861d26422e9fcec0bd9877e9a" Feb 14 11:38:24 crc kubenswrapper[4904]: I0214 11:38:24.047392 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-msvcb"] Feb 14 11:38:24 crc kubenswrapper[4904]: I0214 11:38:24.057987 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-sjsmr"] Feb 14 11:38:24 crc kubenswrapper[4904]: I0214 11:38:24.066371 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-94bx7"] Feb 14 11:38:24 crc kubenswrapper[4904]: I0214 11:38:24.073117 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-94bx7"] Feb 14 11:38:24 crc kubenswrapper[4904]: I0214 11:38:24.080130 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-sjsmr"] Feb 14 11:38:24 crc kubenswrapper[4904]: I0214 11:38:24.087253 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-msvcb"] Feb 14 11:38:25 crc kubenswrapper[4904]: I0214 11:38:25.029978 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-9989-account-create-update-f5ft6"] Feb 14 11:38:25 crc kubenswrapper[4904]: I0214 11:38:25.037612 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-3e19-account-create-update-bxrv4"] Feb 14 11:38:25 crc kubenswrapper[4904]: I0214 11:38:25.045068 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-65a0-account-create-update-lz527"] Feb 14 11:38:25 crc kubenswrapper[4904]: I0214 11:38:25.052299 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-65a0-account-create-update-lz527"] Feb 14 11:38:25 crc kubenswrapper[4904]: I0214 11:38:25.077925 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-3e19-account-create-update-bxrv4"] Feb 14 11:38:25 crc kubenswrapper[4904]: I0214 11:38:25.092561 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-9989-account-create-update-f5ft6"] Feb 14 11:38:25 crc kubenswrapper[4904]: I0214 11:38:25.857301 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="007dfd5a-635a-42ff-b733-1022fe21b6e4" path="/var/lib/kubelet/pods/007dfd5a-635a-42ff-b733-1022fe21b6e4/volumes" Feb 14 11:38:25 crc kubenswrapper[4904]: I0214 11:38:25.858746 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="193a644c-698e-4ad6-a36d-8c19c8ef0d0a" path="/var/lib/kubelet/pods/193a644c-698e-4ad6-a36d-8c19c8ef0d0a/volumes" Feb 14 11:38:25 crc kubenswrapper[4904]: I0214 11:38:25.860053 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d63d7a3-5e12-428e-a9dc-705b28836ca0" path="/var/lib/kubelet/pods/1d63d7a3-5e12-428e-a9dc-705b28836ca0/volumes" Feb 14 11:38:25 crc kubenswrapper[4904]: I0214 11:38:25.861339 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43b09009-a1b8-4f5f-8875-8a01b9672adb" path="/var/lib/kubelet/pods/43b09009-a1b8-4f5f-8875-8a01b9672adb/volumes" Feb 14 11:38:25 crc kubenswrapper[4904]: I0214 11:38:25.863975 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60971f9e-6866-4f58-9686-0eb77268650a" path="/var/lib/kubelet/pods/60971f9e-6866-4f58-9686-0eb77268650a/volumes" Feb 14 11:38:25 crc kubenswrapper[4904]: I0214 11:38:25.865563 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8121d8f-c473-4818-80f2-1335e44a7955" path="/var/lib/kubelet/pods/e8121d8f-c473-4818-80f2-1335e44a7955/volumes" Feb 14 11:38:29 crc kubenswrapper[4904]: I0214 11:38:29.842062 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:38:29 crc kubenswrapper[4904]: E0214 11:38:29.842785 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:38:44 crc kubenswrapper[4904]: I0214 11:38:44.836185 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:38:44 crc kubenswrapper[4904]: E0214 11:38:44.837046 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:38:54 crc kubenswrapper[4904]: I0214 11:38:54.392968 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xdnmn"] Feb 14 11:38:54 crc kubenswrapper[4904]: I0214 11:38:54.402145 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xdnmn"] Feb 14 11:38:55 crc kubenswrapper[4904]: I0214 11:38:55.849466 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6971dbc2-c11e-44d1-b63a-e0487f339183" path="/var/lib/kubelet/pods/6971dbc2-c11e-44d1-b63a-e0487f339183/volumes" Feb 14 11:38:57 crc kubenswrapper[4904]: I0214 11:38:57.798528 4904 generic.go:334] "Generic (PLEG): container finished" podID="80fcf5ac-b576-4551-8478-d411430ad030" containerID="747c70f7a52b504c3d376ef4f28838450a38749ea1c4b9be66b060f37274963c" exitCode=2 Feb 14 11:38:57 crc kubenswrapper[4904]: I0214 11:38:57.798568 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" event={"ID":"80fcf5ac-b576-4551-8478-d411430ad030","Type":"ContainerDied","Data":"747c70f7a52b504c3d376ef4f28838450a38749ea1c4b9be66b060f37274963c"} Feb 14 11:38:59 crc kubenswrapper[4904]: I0214 11:38:59.339070 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" Feb 14 11:38:59 crc kubenswrapper[4904]: I0214 11:38:59.434502 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/80fcf5ac-b576-4551-8478-d411430ad030-ssh-key-openstack-edpm-ipam\") pod \"80fcf5ac-b576-4551-8478-d411430ad030\" (UID: \"80fcf5ac-b576-4551-8478-d411430ad030\") " Feb 14 11:38:59 crc kubenswrapper[4904]: I0214 11:38:59.434856 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80fcf5ac-b576-4551-8478-d411430ad030-repo-setup-combined-ca-bundle\") pod \"80fcf5ac-b576-4551-8478-d411430ad030\" (UID: \"80fcf5ac-b576-4551-8478-d411430ad030\") " Feb 14 11:38:59 crc kubenswrapper[4904]: I0214 11:38:59.434885 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4gpv\" (UniqueName: \"kubernetes.io/projected/80fcf5ac-b576-4551-8478-d411430ad030-kube-api-access-q4gpv\") pod \"80fcf5ac-b576-4551-8478-d411430ad030\" (UID: \"80fcf5ac-b576-4551-8478-d411430ad030\") " Feb 14 11:38:59 crc kubenswrapper[4904]: I0214 11:38:59.434923 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80fcf5ac-b576-4551-8478-d411430ad030-inventory\") pod \"80fcf5ac-b576-4551-8478-d411430ad030\" (UID: \"80fcf5ac-b576-4551-8478-d411430ad030\") " Feb 14 11:38:59 crc kubenswrapper[4904]: I0214 11:38:59.442007 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80fcf5ac-b576-4551-8478-d411430ad030-kube-api-access-q4gpv" (OuterVolumeSpecName: "kube-api-access-q4gpv") pod "80fcf5ac-b576-4551-8478-d411430ad030" (UID: "80fcf5ac-b576-4551-8478-d411430ad030"). InnerVolumeSpecName "kube-api-access-q4gpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:38:59 crc kubenswrapper[4904]: I0214 11:38:59.446326 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80fcf5ac-b576-4551-8478-d411430ad030-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "80fcf5ac-b576-4551-8478-d411430ad030" (UID: "80fcf5ac-b576-4551-8478-d411430ad030"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:38:59 crc kubenswrapper[4904]: I0214 11:38:59.471051 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80fcf5ac-b576-4551-8478-d411430ad030-inventory" (OuterVolumeSpecName: "inventory") pod "80fcf5ac-b576-4551-8478-d411430ad030" (UID: "80fcf5ac-b576-4551-8478-d411430ad030"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:38:59 crc kubenswrapper[4904]: I0214 11:38:59.474800 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80fcf5ac-b576-4551-8478-d411430ad030-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "80fcf5ac-b576-4551-8478-d411430ad030" (UID: "80fcf5ac-b576-4551-8478-d411430ad030"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:38:59 crc kubenswrapper[4904]: I0214 11:38:59.537191 4904 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/80fcf5ac-b576-4551-8478-d411430ad030-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 14 11:38:59 crc kubenswrapper[4904]: I0214 11:38:59.537227 4904 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80fcf5ac-b576-4551-8478-d411430ad030-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:38:59 crc kubenswrapper[4904]: I0214 11:38:59.537237 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4gpv\" (UniqueName: \"kubernetes.io/projected/80fcf5ac-b576-4551-8478-d411430ad030-kube-api-access-q4gpv\") on node \"crc\" DevicePath \"\"" Feb 14 11:38:59 crc kubenswrapper[4904]: I0214 11:38:59.537248 4904 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80fcf5ac-b576-4551-8478-d411430ad030-inventory\") on node \"crc\" DevicePath \"\"" Feb 14 11:38:59 crc kubenswrapper[4904]: I0214 11:38:59.813601 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" event={"ID":"80fcf5ac-b576-4551-8478-d411430ad030","Type":"ContainerDied","Data":"fa09f6f67314090aa6b0e6ee4033fe52073cb231299066e8811651b4ecd2de89"} Feb 14 11:38:59 crc kubenswrapper[4904]: I0214 11:38:59.813640 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa09f6f67314090aa6b0e6ee4033fe52073cb231299066e8811651b4ecd2de89" Feb 14 11:38:59 crc kubenswrapper[4904]: I0214 11:38:59.813683 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj" Feb 14 11:38:59 crc kubenswrapper[4904]: I0214 11:38:59.842773 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:38:59 crc kubenswrapper[4904]: E0214 11:38:59.843120 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:39:13 crc kubenswrapper[4904]: I0214 11:39:13.836770 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:39:13 crc kubenswrapper[4904]: E0214 11:39:13.837537 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:39:14 crc kubenswrapper[4904]: I0214 11:39:14.030819 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-98lq8"] Feb 14 11:39:14 crc kubenswrapper[4904]: I0214 11:39:14.040740 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-98lq8"] Feb 14 11:39:15 crc kubenswrapper[4904]: I0214 11:39:15.027646 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-nvrps"] Feb 14 11:39:15 crc kubenswrapper[4904]: I0214 11:39:15.035584 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-nvrps"] Feb 14 11:39:15 crc kubenswrapper[4904]: I0214 11:39:15.110011 4904 scope.go:117] "RemoveContainer" containerID="849751e6d1a258192b1d4909642f031efd3aabe21275e10c111c340def2869e0" Feb 14 11:39:15 crc kubenswrapper[4904]: I0214 11:39:15.138890 4904 scope.go:117] "RemoveContainer" containerID="7bbe956539df1f0dd3029213f6239b45e4c549f580c63afb238146cbf2714a48" Feb 14 11:39:15 crc kubenswrapper[4904]: I0214 11:39:15.203370 4904 scope.go:117] "RemoveContainer" containerID="f5d2afde37cb14f49005661ff25620cc8f7cb4912b6cff6c69c1cd7bad02af86" Feb 14 11:39:15 crc kubenswrapper[4904]: I0214 11:39:15.234375 4904 scope.go:117] "RemoveContainer" containerID="5b00d53d5a2e3c969cf00ca830a19f3265c27bd06c9f6bd11562172db8231133" Feb 14 11:39:15 crc kubenswrapper[4904]: I0214 11:39:15.272997 4904 scope.go:117] "RemoveContainer" containerID="5617a814dc9d8473995bd32f3185aa1c209505034d59ccfbb4cf88e183c7b9bb" Feb 14 11:39:15 crc kubenswrapper[4904]: I0214 11:39:15.310197 4904 scope.go:117] "RemoveContainer" containerID="38d86ea826aaa0b0c338baa908f8e2eec8ab8645a24eb5da05a73e1148880428" Feb 14 11:39:15 crc kubenswrapper[4904]: I0214 11:39:15.361682 4904 scope.go:117] "RemoveContainer" containerID="a7c9eb679c798b42518db4d0d8280720eaad17b33ac18c43dd6196669663e7a4" Feb 14 11:39:15 crc kubenswrapper[4904]: I0214 11:39:15.847875 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3011b1dc-9d5e-4c28-bdf6-de2199704f42" path="/var/lib/kubelet/pods/3011b1dc-9d5e-4c28-bdf6-de2199704f42/volumes" Feb 14 11:39:15 crc kubenswrapper[4904]: I0214 11:39:15.848636 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="806d3bba-5958-4939-90a5-ae94621e1c5a" path="/var/lib/kubelet/pods/806d3bba-5958-4939-90a5-ae94621e1c5a/volumes" Feb 14 11:39:24 crc kubenswrapper[4904]: I0214 11:39:24.837292 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:39:24 crc kubenswrapper[4904]: E0214 11:39:24.838047 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:39:35 crc kubenswrapper[4904]: I0214 11:39:35.836734 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:39:35 crc kubenswrapper[4904]: E0214 11:39:35.837493 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.033850 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj"] Feb 14 11:39:37 crc kubenswrapper[4904]: E0214 11:39:37.034495 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92dc2e95-c354-4388-9e09-2470bfce8e34" containerName="extract-content" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.034507 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="92dc2e95-c354-4388-9e09-2470bfce8e34" containerName="extract-content" Feb 14 11:39:37 crc kubenswrapper[4904]: E0214 11:39:37.034523 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92dc2e95-c354-4388-9e09-2470bfce8e34" containerName="registry-server" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.034530 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="92dc2e95-c354-4388-9e09-2470bfce8e34" containerName="registry-server" Feb 14 11:39:37 crc kubenswrapper[4904]: E0214 11:39:37.034544 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0aea853-283c-4da1-9dfb-f47bd4ed1117" containerName="registry-server" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.034550 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0aea853-283c-4da1-9dfb-f47bd4ed1117" containerName="registry-server" Feb 14 11:39:37 crc kubenswrapper[4904]: E0214 11:39:37.034556 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0aea853-283c-4da1-9dfb-f47bd4ed1117" containerName="extract-content" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.034562 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0aea853-283c-4da1-9dfb-f47bd4ed1117" containerName="extract-content" Feb 14 11:39:37 crc kubenswrapper[4904]: E0214 11:39:37.034571 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0aea853-283c-4da1-9dfb-f47bd4ed1117" containerName="extract-utilities" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.034577 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0aea853-283c-4da1-9dfb-f47bd4ed1117" containerName="extract-utilities" Feb 14 11:39:37 crc kubenswrapper[4904]: E0214 11:39:37.034587 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92dc2e95-c354-4388-9e09-2470bfce8e34" containerName="extract-utilities" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.034593 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="92dc2e95-c354-4388-9e09-2470bfce8e34" containerName="extract-utilities" Feb 14 11:39:37 crc kubenswrapper[4904]: E0214 11:39:37.034607 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80fcf5ac-b576-4551-8478-d411430ad030" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.034613 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="80fcf5ac-b576-4551-8478-d411430ad030" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.034785 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="80fcf5ac-b576-4551-8478-d411430ad030" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.034801 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="92dc2e95-c354-4388-9e09-2470bfce8e34" containerName="registry-server" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.034809 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0aea853-283c-4da1-9dfb-f47bd4ed1117" containerName="registry-server" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.035468 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.038066 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.039182 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-48rnz" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.039919 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.043106 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.052584 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj"] Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.131394 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj\" (UID: \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.131621 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj\" (UID: \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.131693 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmjxq\" (UniqueName: \"kubernetes.io/projected/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-kube-api-access-mmjxq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj\" (UID: \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.131737 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj\" (UID: \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.232933 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj\" (UID: \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.232993 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmjxq\" (UniqueName: \"kubernetes.io/projected/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-kube-api-access-mmjxq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj\" (UID: \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.233019 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj\" (UID: \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.233052 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj\" (UID: \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.238612 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj\" (UID: \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.241634 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj\" (UID: \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.242827 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj\" (UID: \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.251687 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmjxq\" (UniqueName: \"kubernetes.io/projected/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-kube-api-access-mmjxq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj\" (UID: \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.362447 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" Feb 14 11:39:37 crc kubenswrapper[4904]: I0214 11:39:37.859662 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj"] Feb 14 11:39:38 crc kubenswrapper[4904]: I0214 11:39:38.121184 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" event={"ID":"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6","Type":"ContainerStarted","Data":"38ea72e8edea1d88df793a773936e43807490d960a8d13906cd530db7aafb897"} Feb 14 11:39:39 crc kubenswrapper[4904]: I0214 11:39:39.129358 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" event={"ID":"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6","Type":"ContainerStarted","Data":"ff6a1082a3b13f91f05388de46187529da853267e96357f44d90a2be875987fa"} Feb 14 11:39:39 crc kubenswrapper[4904]: I0214 11:39:39.145386 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" podStartSLOduration=1.229700258 podStartE2EDuration="2.145364661s" podCreationTimestamp="2026-02-14 11:39:37 +0000 UTC" firstStartedPulling="2026-02-14 11:39:37.875296712 +0000 UTC m=+1768.688061373" lastFinishedPulling="2026-02-14 11:39:38.790961115 +0000 UTC m=+1769.603725776" observedRunningTime="2026-02-14 11:39:39.141235578 +0000 UTC m=+1769.954000239" watchObservedRunningTime="2026-02-14 11:39:39.145364661 +0000 UTC m=+1769.958129322" Feb 14 11:39:50 crc kubenswrapper[4904]: I0214 11:39:50.837147 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:39:50 crc kubenswrapper[4904]: E0214 11:39:50.838614 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:39:58 crc kubenswrapper[4904]: I0214 11:39:58.039819 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-4kbtk"] Feb 14 11:39:58 crc kubenswrapper[4904]: I0214 11:39:58.051361 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-4kbtk"] Feb 14 11:39:59 crc kubenswrapper[4904]: I0214 11:39:59.850782 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29681a87-9c4c-47fd-9834-c3cf2ebd79d3" path="/var/lib/kubelet/pods/29681a87-9c4c-47fd-9834-c3cf2ebd79d3/volumes" Feb 14 11:40:05 crc kubenswrapper[4904]: I0214 11:40:05.836930 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:40:05 crc kubenswrapper[4904]: E0214 11:40:05.838958 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:40:15 crc kubenswrapper[4904]: I0214 11:40:15.509019 4904 scope.go:117] "RemoveContainer" containerID="be4ad0665d650aa5adb7d3d177de2d9743d103e82015f090cd8312b792aa6e44" Feb 14 11:40:15 crc kubenswrapper[4904]: I0214 11:40:15.541772 4904 scope.go:117] "RemoveContainer" containerID="2ceedd3fe499939062c581b036d4eea72ba156e02350e396a4a38a7e2918a2bd" Feb 14 11:40:15 crc kubenswrapper[4904]: I0214 11:40:15.578351 4904 scope.go:117] "RemoveContainer" containerID="2b034addfe9ba1b39e700ac2917b72a677f2e05b1ddc127dd0c3ed3a7f26a1c8" Feb 14 11:40:19 crc kubenswrapper[4904]: I0214 11:40:19.845719 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:40:19 crc kubenswrapper[4904]: E0214 11:40:19.846760 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:40:31 crc kubenswrapper[4904]: I0214 11:40:31.836549 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:40:31 crc kubenswrapper[4904]: E0214 11:40:31.838173 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:40:42 crc kubenswrapper[4904]: I0214 11:40:42.836576 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:40:42 crc kubenswrapper[4904]: E0214 11:40:42.837415 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:40:51 crc kubenswrapper[4904]: I0214 11:40:51.101318 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zkt4t"] Feb 14 11:40:51 crc kubenswrapper[4904]: I0214 11:40:51.106850 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zkt4t" Feb 14 11:40:51 crc kubenswrapper[4904]: I0214 11:40:51.109667 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zkt4t"] Feb 14 11:40:51 crc kubenswrapper[4904]: I0214 11:40:51.158272 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a96f426c-8576-4f0b-8d83-f554dece8528-utilities\") pod \"community-operators-zkt4t\" (UID: \"a96f426c-8576-4f0b-8d83-f554dece8528\") " pod="openshift-marketplace/community-operators-zkt4t" Feb 14 11:40:51 crc kubenswrapper[4904]: I0214 11:40:51.158451 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a96f426c-8576-4f0b-8d83-f554dece8528-catalog-content\") pod \"community-operators-zkt4t\" (UID: \"a96f426c-8576-4f0b-8d83-f554dece8528\") " pod="openshift-marketplace/community-operators-zkt4t" Feb 14 11:40:51 crc kubenswrapper[4904]: I0214 11:40:51.158493 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bln8d\" (UniqueName: \"kubernetes.io/projected/a96f426c-8576-4f0b-8d83-f554dece8528-kube-api-access-bln8d\") pod \"community-operators-zkt4t\" (UID: \"a96f426c-8576-4f0b-8d83-f554dece8528\") " pod="openshift-marketplace/community-operators-zkt4t" Feb 14 11:40:51 crc kubenswrapper[4904]: I0214 11:40:51.259981 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bln8d\" (UniqueName: \"kubernetes.io/projected/a96f426c-8576-4f0b-8d83-f554dece8528-kube-api-access-bln8d\") pod \"community-operators-zkt4t\" (UID: \"a96f426c-8576-4f0b-8d83-f554dece8528\") " pod="openshift-marketplace/community-operators-zkt4t" Feb 14 11:40:51 crc kubenswrapper[4904]: I0214 11:40:51.260389 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a96f426c-8576-4f0b-8d83-f554dece8528-utilities\") pod \"community-operators-zkt4t\" (UID: \"a96f426c-8576-4f0b-8d83-f554dece8528\") " pod="openshift-marketplace/community-operators-zkt4t" Feb 14 11:40:51 crc kubenswrapper[4904]: I0214 11:40:51.260671 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a96f426c-8576-4f0b-8d83-f554dece8528-catalog-content\") pod \"community-operators-zkt4t\" (UID: \"a96f426c-8576-4f0b-8d83-f554dece8528\") " pod="openshift-marketplace/community-operators-zkt4t" Feb 14 11:40:51 crc kubenswrapper[4904]: I0214 11:40:51.260903 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a96f426c-8576-4f0b-8d83-f554dece8528-utilities\") pod \"community-operators-zkt4t\" (UID: \"a96f426c-8576-4f0b-8d83-f554dece8528\") " pod="openshift-marketplace/community-operators-zkt4t" Feb 14 11:40:51 crc kubenswrapper[4904]: I0214 11:40:51.261212 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a96f426c-8576-4f0b-8d83-f554dece8528-catalog-content\") pod \"community-operators-zkt4t\" (UID: \"a96f426c-8576-4f0b-8d83-f554dece8528\") " pod="openshift-marketplace/community-operators-zkt4t" Feb 14 11:40:51 crc kubenswrapper[4904]: I0214 11:40:51.286806 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bln8d\" (UniqueName: \"kubernetes.io/projected/a96f426c-8576-4f0b-8d83-f554dece8528-kube-api-access-bln8d\") pod \"community-operators-zkt4t\" (UID: \"a96f426c-8576-4f0b-8d83-f554dece8528\") " pod="openshift-marketplace/community-operators-zkt4t" Feb 14 11:40:51 crc kubenswrapper[4904]: I0214 11:40:51.426674 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zkt4t" Feb 14 11:40:52 crc kubenswrapper[4904]: I0214 11:40:52.066861 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zkt4t"] Feb 14 11:40:52 crc kubenswrapper[4904]: I0214 11:40:52.795160 4904 generic.go:334] "Generic (PLEG): container finished" podID="a96f426c-8576-4f0b-8d83-f554dece8528" containerID="af4e877908b433ce494de0db4663b2739d6d947ee6e7f26a79b28d24afffe8cc" exitCode=0 Feb 14 11:40:52 crc kubenswrapper[4904]: I0214 11:40:52.795207 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkt4t" event={"ID":"a96f426c-8576-4f0b-8d83-f554dece8528","Type":"ContainerDied","Data":"af4e877908b433ce494de0db4663b2739d6d947ee6e7f26a79b28d24afffe8cc"} Feb 14 11:40:52 crc kubenswrapper[4904]: I0214 11:40:52.795237 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkt4t" event={"ID":"a96f426c-8576-4f0b-8d83-f554dece8528","Type":"ContainerStarted","Data":"3dc8f1a514d594c35a31045aafbdd60ee3b5afd558adc5d5d8b8889a56da294a"} Feb 14 11:40:53 crc kubenswrapper[4904]: I0214 11:40:53.530950 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mzsr8"] Feb 14 11:40:53 crc kubenswrapper[4904]: I0214 11:40:53.533162 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mzsr8" Feb 14 11:40:53 crc kubenswrapper[4904]: I0214 11:40:53.594999 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzsr8"] Feb 14 11:40:53 crc kubenswrapper[4904]: I0214 11:40:53.602555 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fa860db-2aad-4bdd-a920-390bc4701f0f-catalog-content\") pod \"redhat-marketplace-mzsr8\" (UID: \"1fa860db-2aad-4bdd-a920-390bc4701f0f\") " pod="openshift-marketplace/redhat-marketplace-mzsr8" Feb 14 11:40:53 crc kubenswrapper[4904]: I0214 11:40:53.603228 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkgbp\" (UniqueName: \"kubernetes.io/projected/1fa860db-2aad-4bdd-a920-390bc4701f0f-kube-api-access-zkgbp\") pod \"redhat-marketplace-mzsr8\" (UID: \"1fa860db-2aad-4bdd-a920-390bc4701f0f\") " pod="openshift-marketplace/redhat-marketplace-mzsr8" Feb 14 11:40:53 crc kubenswrapper[4904]: I0214 11:40:53.604015 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fa860db-2aad-4bdd-a920-390bc4701f0f-utilities\") pod \"redhat-marketplace-mzsr8\" (UID: \"1fa860db-2aad-4bdd-a920-390bc4701f0f\") " pod="openshift-marketplace/redhat-marketplace-mzsr8" Feb 14 11:40:53 crc kubenswrapper[4904]: I0214 11:40:53.705948 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fa860db-2aad-4bdd-a920-390bc4701f0f-catalog-content\") pod \"redhat-marketplace-mzsr8\" (UID: \"1fa860db-2aad-4bdd-a920-390bc4701f0f\") " pod="openshift-marketplace/redhat-marketplace-mzsr8" Feb 14 11:40:53 crc kubenswrapper[4904]: I0214 11:40:53.706037 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkgbp\" (UniqueName: \"kubernetes.io/projected/1fa860db-2aad-4bdd-a920-390bc4701f0f-kube-api-access-zkgbp\") pod \"redhat-marketplace-mzsr8\" (UID: \"1fa860db-2aad-4bdd-a920-390bc4701f0f\") " pod="openshift-marketplace/redhat-marketplace-mzsr8" Feb 14 11:40:53 crc kubenswrapper[4904]: I0214 11:40:53.706078 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fa860db-2aad-4bdd-a920-390bc4701f0f-utilities\") pod \"redhat-marketplace-mzsr8\" (UID: \"1fa860db-2aad-4bdd-a920-390bc4701f0f\") " pod="openshift-marketplace/redhat-marketplace-mzsr8" Feb 14 11:40:53 crc kubenswrapper[4904]: I0214 11:40:53.706506 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fa860db-2aad-4bdd-a920-390bc4701f0f-utilities\") pod \"redhat-marketplace-mzsr8\" (UID: \"1fa860db-2aad-4bdd-a920-390bc4701f0f\") " pod="openshift-marketplace/redhat-marketplace-mzsr8" Feb 14 11:40:53 crc kubenswrapper[4904]: I0214 11:40:53.706701 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fa860db-2aad-4bdd-a920-390bc4701f0f-catalog-content\") pod \"redhat-marketplace-mzsr8\" (UID: \"1fa860db-2aad-4bdd-a920-390bc4701f0f\") " pod="openshift-marketplace/redhat-marketplace-mzsr8" Feb 14 11:40:53 crc kubenswrapper[4904]: I0214 11:40:53.726554 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkgbp\" (UniqueName: \"kubernetes.io/projected/1fa860db-2aad-4bdd-a920-390bc4701f0f-kube-api-access-zkgbp\") pod \"redhat-marketplace-mzsr8\" (UID: \"1fa860db-2aad-4bdd-a920-390bc4701f0f\") " pod="openshift-marketplace/redhat-marketplace-mzsr8" Feb 14 11:40:53 crc kubenswrapper[4904]: I0214 11:40:53.804128 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkt4t" event={"ID":"a96f426c-8576-4f0b-8d83-f554dece8528","Type":"ContainerStarted","Data":"faadfd8a991b4d1b664ce0dfca2c33b216366945723d61370aa460e5e66e14e0"} Feb 14 11:40:53 crc kubenswrapper[4904]: I0214 11:40:53.836636 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:40:53 crc kubenswrapper[4904]: I0214 11:40:53.903954 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mzsr8" Feb 14 11:40:54 crc kubenswrapper[4904]: I0214 11:40:54.447624 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzsr8"] Feb 14 11:40:54 crc kubenswrapper[4904]: I0214 11:40:54.813359 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"36065fedbbfd5007e91cc94e16b80606571b96f7d01d643f9259259819fe5f66"} Feb 14 11:40:54 crc kubenswrapper[4904]: I0214 11:40:54.814807 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzsr8" event={"ID":"1fa860db-2aad-4bdd-a920-390bc4701f0f","Type":"ContainerDied","Data":"6e0e9ef14ca959e94259e0ffe0cff873e397ef03974a907d076f39b2c9a119f4"} Feb 14 11:40:54 crc kubenswrapper[4904]: I0214 11:40:54.814814 4904 generic.go:334] "Generic (PLEG): container finished" podID="1fa860db-2aad-4bdd-a920-390bc4701f0f" containerID="6e0e9ef14ca959e94259e0ffe0cff873e397ef03974a907d076f39b2c9a119f4" exitCode=0 Feb 14 11:40:54 crc kubenswrapper[4904]: I0214 11:40:54.815050 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzsr8" event={"ID":"1fa860db-2aad-4bdd-a920-390bc4701f0f","Type":"ContainerStarted","Data":"01dd16a9846350e47d454fde0cf4af6c7e23c3c6141d7909b95232d78f4f75b5"} Feb 14 11:40:55 crc kubenswrapper[4904]: I0214 11:40:55.824530 4904 generic.go:334] "Generic (PLEG): container finished" podID="a96f426c-8576-4f0b-8d83-f554dece8528" containerID="faadfd8a991b4d1b664ce0dfca2c33b216366945723d61370aa460e5e66e14e0" exitCode=0 Feb 14 11:40:55 crc kubenswrapper[4904]: I0214 11:40:55.824608 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkt4t" event={"ID":"a96f426c-8576-4f0b-8d83-f554dece8528","Type":"ContainerDied","Data":"faadfd8a991b4d1b664ce0dfca2c33b216366945723d61370aa460e5e66e14e0"} Feb 14 11:40:55 crc kubenswrapper[4904]: I0214 11:40:55.830886 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzsr8" event={"ID":"1fa860db-2aad-4bdd-a920-390bc4701f0f","Type":"ContainerStarted","Data":"eb536480ede3d5c2124f3c1899e875214c8235c2fe04933f95415cff5f2a7dc7"} Feb 14 11:40:56 crc kubenswrapper[4904]: I0214 11:40:56.842568 4904 generic.go:334] "Generic (PLEG): container finished" podID="1fa860db-2aad-4bdd-a920-390bc4701f0f" containerID="eb536480ede3d5c2124f3c1899e875214c8235c2fe04933f95415cff5f2a7dc7" exitCode=0 Feb 14 11:40:56 crc kubenswrapper[4904]: I0214 11:40:56.843555 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzsr8" event={"ID":"1fa860db-2aad-4bdd-a920-390bc4701f0f","Type":"ContainerDied","Data":"eb536480ede3d5c2124f3c1899e875214c8235c2fe04933f95415cff5f2a7dc7"} Feb 14 11:40:56 crc kubenswrapper[4904]: I0214 11:40:56.849852 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkt4t" event={"ID":"a96f426c-8576-4f0b-8d83-f554dece8528","Type":"ContainerStarted","Data":"cf73530e0606b3feb2c9de310817c5bea835de7a7bc1e58d6ded621b2be185f0"} Feb 14 11:40:56 crc kubenswrapper[4904]: I0214 11:40:56.890312 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zkt4t" podStartSLOduration=2.442699884 podStartE2EDuration="5.89029285s" podCreationTimestamp="2026-02-14 11:40:51 +0000 UTC" firstStartedPulling="2026-02-14 11:40:52.797126606 +0000 UTC m=+1843.609891267" lastFinishedPulling="2026-02-14 11:40:56.244719572 +0000 UTC m=+1847.057484233" observedRunningTime="2026-02-14 11:40:56.882913688 +0000 UTC m=+1847.695678349" watchObservedRunningTime="2026-02-14 11:40:56.89029285 +0000 UTC m=+1847.703057521" Feb 14 11:40:57 crc kubenswrapper[4904]: I0214 11:40:57.860628 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzsr8" event={"ID":"1fa860db-2aad-4bdd-a920-390bc4701f0f","Type":"ContainerStarted","Data":"0bb4ad306413108c6d3ac874261aade422dfe0ee7216648679726d6272c0c32b"} Feb 14 11:41:01 crc kubenswrapper[4904]: I0214 11:41:01.426898 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zkt4t" Feb 14 11:41:01 crc kubenswrapper[4904]: I0214 11:41:01.427440 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zkt4t" Feb 14 11:41:01 crc kubenswrapper[4904]: I0214 11:41:01.493495 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zkt4t" Feb 14 11:41:01 crc kubenswrapper[4904]: I0214 11:41:01.513911 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mzsr8" podStartSLOduration=6.054083371 podStartE2EDuration="8.513891067s" podCreationTimestamp="2026-02-14 11:40:53 +0000 UTC" firstStartedPulling="2026-02-14 11:40:54.81660513 +0000 UTC m=+1845.629369791" lastFinishedPulling="2026-02-14 11:40:57.276412826 +0000 UTC m=+1848.089177487" observedRunningTime="2026-02-14 11:40:57.886423009 +0000 UTC m=+1848.699187670" watchObservedRunningTime="2026-02-14 11:41:01.513891067 +0000 UTC m=+1852.326655738" Feb 14 11:41:01 crc kubenswrapper[4904]: I0214 11:41:01.944074 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zkt4t" Feb 14 11:41:03 crc kubenswrapper[4904]: I0214 11:41:03.093525 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zkt4t"] Feb 14 11:41:03 crc kubenswrapper[4904]: I0214 11:41:03.904053 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mzsr8" Feb 14 11:41:03 crc kubenswrapper[4904]: I0214 11:41:03.905514 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mzsr8" Feb 14 11:41:03 crc kubenswrapper[4904]: I0214 11:41:03.904778 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zkt4t" podUID="a96f426c-8576-4f0b-8d83-f554dece8528" containerName="registry-server" containerID="cri-o://cf73530e0606b3feb2c9de310817c5bea835de7a7bc1e58d6ded621b2be185f0" gracePeriod=2 Feb 14 11:41:03 crc kubenswrapper[4904]: I0214 11:41:03.954956 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mzsr8" Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.343289 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zkt4t" Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.417862 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a96f426c-8576-4f0b-8d83-f554dece8528-catalog-content\") pod \"a96f426c-8576-4f0b-8d83-f554dece8528\" (UID: \"a96f426c-8576-4f0b-8d83-f554dece8528\") " Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.418350 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a96f426c-8576-4f0b-8d83-f554dece8528-utilities\") pod \"a96f426c-8576-4f0b-8d83-f554dece8528\" (UID: \"a96f426c-8576-4f0b-8d83-f554dece8528\") " Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.418464 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bln8d\" (UniqueName: \"kubernetes.io/projected/a96f426c-8576-4f0b-8d83-f554dece8528-kube-api-access-bln8d\") pod \"a96f426c-8576-4f0b-8d83-f554dece8528\" (UID: \"a96f426c-8576-4f0b-8d83-f554dece8528\") " Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.419102 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a96f426c-8576-4f0b-8d83-f554dece8528-utilities" (OuterVolumeSpecName: "utilities") pod "a96f426c-8576-4f0b-8d83-f554dece8528" (UID: "a96f426c-8576-4f0b-8d83-f554dece8528"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.419680 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a96f426c-8576-4f0b-8d83-f554dece8528-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.423912 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a96f426c-8576-4f0b-8d83-f554dece8528-kube-api-access-bln8d" (OuterVolumeSpecName: "kube-api-access-bln8d") pod "a96f426c-8576-4f0b-8d83-f554dece8528" (UID: "a96f426c-8576-4f0b-8d83-f554dece8528"). InnerVolumeSpecName "kube-api-access-bln8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.472408 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a96f426c-8576-4f0b-8d83-f554dece8528-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a96f426c-8576-4f0b-8d83-f554dece8528" (UID: "a96f426c-8576-4f0b-8d83-f554dece8528"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.521385 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bln8d\" (UniqueName: \"kubernetes.io/projected/a96f426c-8576-4f0b-8d83-f554dece8528-kube-api-access-bln8d\") on node \"crc\" DevicePath \"\"" Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.521459 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a96f426c-8576-4f0b-8d83-f554dece8528-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.913683 4904 generic.go:334] "Generic (PLEG): container finished" podID="a96f426c-8576-4f0b-8d83-f554dece8528" containerID="cf73530e0606b3feb2c9de310817c5bea835de7a7bc1e58d6ded621b2be185f0" exitCode=0 Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.913762 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkt4t" event={"ID":"a96f426c-8576-4f0b-8d83-f554dece8528","Type":"ContainerDied","Data":"cf73530e0606b3feb2c9de310817c5bea835de7a7bc1e58d6ded621b2be185f0"} Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.913787 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zkt4t" Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.913796 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zkt4t" event={"ID":"a96f426c-8576-4f0b-8d83-f554dece8528","Type":"ContainerDied","Data":"3dc8f1a514d594c35a31045aafbdd60ee3b5afd558adc5d5d8b8889a56da294a"} Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.913819 4904 scope.go:117] "RemoveContainer" containerID="cf73530e0606b3feb2c9de310817c5bea835de7a7bc1e58d6ded621b2be185f0" Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.933107 4904 scope.go:117] "RemoveContainer" containerID="faadfd8a991b4d1b664ce0dfca2c33b216366945723d61370aa460e5e66e14e0" Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.954886 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zkt4t"] Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.961916 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zkt4t"] Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.966543 4904 scope.go:117] "RemoveContainer" containerID="af4e877908b433ce494de0db4663b2739d6d947ee6e7f26a79b28d24afffe8cc" Feb 14 11:41:04 crc kubenswrapper[4904]: I0214 11:41:04.976318 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mzsr8" Feb 14 11:41:05 crc kubenswrapper[4904]: I0214 11:41:05.033035 4904 scope.go:117] "RemoveContainer" containerID="cf73530e0606b3feb2c9de310817c5bea835de7a7bc1e58d6ded621b2be185f0" Feb 14 11:41:05 crc kubenswrapper[4904]: E0214 11:41:05.033491 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf73530e0606b3feb2c9de310817c5bea835de7a7bc1e58d6ded621b2be185f0\": container with ID starting with cf73530e0606b3feb2c9de310817c5bea835de7a7bc1e58d6ded621b2be185f0 not found: ID does not exist" containerID="cf73530e0606b3feb2c9de310817c5bea835de7a7bc1e58d6ded621b2be185f0" Feb 14 11:41:05 crc kubenswrapper[4904]: I0214 11:41:05.033520 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf73530e0606b3feb2c9de310817c5bea835de7a7bc1e58d6ded621b2be185f0"} err="failed to get container status \"cf73530e0606b3feb2c9de310817c5bea835de7a7bc1e58d6ded621b2be185f0\": rpc error: code = NotFound desc = could not find container \"cf73530e0606b3feb2c9de310817c5bea835de7a7bc1e58d6ded621b2be185f0\": container with ID starting with cf73530e0606b3feb2c9de310817c5bea835de7a7bc1e58d6ded621b2be185f0 not found: ID does not exist" Feb 14 11:41:05 crc kubenswrapper[4904]: I0214 11:41:05.033539 4904 scope.go:117] "RemoveContainer" containerID="faadfd8a991b4d1b664ce0dfca2c33b216366945723d61370aa460e5e66e14e0" Feb 14 11:41:05 crc kubenswrapper[4904]: E0214 11:41:05.033884 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"faadfd8a991b4d1b664ce0dfca2c33b216366945723d61370aa460e5e66e14e0\": container with ID starting with faadfd8a991b4d1b664ce0dfca2c33b216366945723d61370aa460e5e66e14e0 not found: ID does not exist" containerID="faadfd8a991b4d1b664ce0dfca2c33b216366945723d61370aa460e5e66e14e0" Feb 14 11:41:05 crc kubenswrapper[4904]: I0214 11:41:05.033903 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faadfd8a991b4d1b664ce0dfca2c33b216366945723d61370aa460e5e66e14e0"} err="failed to get container status \"faadfd8a991b4d1b664ce0dfca2c33b216366945723d61370aa460e5e66e14e0\": rpc error: code = NotFound desc = could not find container \"faadfd8a991b4d1b664ce0dfca2c33b216366945723d61370aa460e5e66e14e0\": container with ID starting with faadfd8a991b4d1b664ce0dfca2c33b216366945723d61370aa460e5e66e14e0 not found: ID does not exist" Feb 14 11:41:05 crc kubenswrapper[4904]: I0214 11:41:05.033916 4904 scope.go:117] "RemoveContainer" containerID="af4e877908b433ce494de0db4663b2739d6d947ee6e7f26a79b28d24afffe8cc" Feb 14 11:41:05 crc kubenswrapper[4904]: E0214 11:41:05.034216 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af4e877908b433ce494de0db4663b2739d6d947ee6e7f26a79b28d24afffe8cc\": container with ID starting with af4e877908b433ce494de0db4663b2739d6d947ee6e7f26a79b28d24afffe8cc not found: ID does not exist" containerID="af4e877908b433ce494de0db4663b2739d6d947ee6e7f26a79b28d24afffe8cc" Feb 14 11:41:05 crc kubenswrapper[4904]: I0214 11:41:05.034237 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af4e877908b433ce494de0db4663b2739d6d947ee6e7f26a79b28d24afffe8cc"} err="failed to get container status \"af4e877908b433ce494de0db4663b2739d6d947ee6e7f26a79b28d24afffe8cc\": rpc error: code = NotFound desc = could not find container \"af4e877908b433ce494de0db4663b2739d6d947ee6e7f26a79b28d24afffe8cc\": container with ID starting with af4e877908b433ce494de0db4663b2739d6d947ee6e7f26a79b28d24afffe8cc not found: ID does not exist" Feb 14 11:41:05 crc kubenswrapper[4904]: I0214 11:41:05.850945 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a96f426c-8576-4f0b-8d83-f554dece8528" path="/var/lib/kubelet/pods/a96f426c-8576-4f0b-8d83-f554dece8528/volumes" Feb 14 11:41:06 crc kubenswrapper[4904]: I0214 11:41:06.289817 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzsr8"] Feb 14 11:41:06 crc kubenswrapper[4904]: I0214 11:41:06.931511 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mzsr8" podUID="1fa860db-2aad-4bdd-a920-390bc4701f0f" containerName="registry-server" containerID="cri-o://0bb4ad306413108c6d3ac874261aade422dfe0ee7216648679726d6272c0c32b" gracePeriod=2 Feb 14 11:41:07 crc kubenswrapper[4904]: I0214 11:41:07.402209 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mzsr8" Feb 14 11:41:07 crc kubenswrapper[4904]: I0214 11:41:07.475362 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkgbp\" (UniqueName: \"kubernetes.io/projected/1fa860db-2aad-4bdd-a920-390bc4701f0f-kube-api-access-zkgbp\") pod \"1fa860db-2aad-4bdd-a920-390bc4701f0f\" (UID: \"1fa860db-2aad-4bdd-a920-390bc4701f0f\") " Feb 14 11:41:07 crc kubenswrapper[4904]: I0214 11:41:07.475433 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fa860db-2aad-4bdd-a920-390bc4701f0f-catalog-content\") pod \"1fa860db-2aad-4bdd-a920-390bc4701f0f\" (UID: \"1fa860db-2aad-4bdd-a920-390bc4701f0f\") " Feb 14 11:41:07 crc kubenswrapper[4904]: I0214 11:41:07.475588 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fa860db-2aad-4bdd-a920-390bc4701f0f-utilities\") pod \"1fa860db-2aad-4bdd-a920-390bc4701f0f\" (UID: \"1fa860db-2aad-4bdd-a920-390bc4701f0f\") " Feb 14 11:41:07 crc kubenswrapper[4904]: I0214 11:41:07.476355 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fa860db-2aad-4bdd-a920-390bc4701f0f-utilities" (OuterVolumeSpecName: "utilities") pod "1fa860db-2aad-4bdd-a920-390bc4701f0f" (UID: "1fa860db-2aad-4bdd-a920-390bc4701f0f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:41:07 crc kubenswrapper[4904]: I0214 11:41:07.480548 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fa860db-2aad-4bdd-a920-390bc4701f0f-kube-api-access-zkgbp" (OuterVolumeSpecName: "kube-api-access-zkgbp") pod "1fa860db-2aad-4bdd-a920-390bc4701f0f" (UID: "1fa860db-2aad-4bdd-a920-390bc4701f0f"). InnerVolumeSpecName "kube-api-access-zkgbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:41:07 crc kubenswrapper[4904]: I0214 11:41:07.500971 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fa860db-2aad-4bdd-a920-390bc4701f0f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1fa860db-2aad-4bdd-a920-390bc4701f0f" (UID: "1fa860db-2aad-4bdd-a920-390bc4701f0f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:41:07 crc kubenswrapper[4904]: I0214 11:41:07.577369 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkgbp\" (UniqueName: \"kubernetes.io/projected/1fa860db-2aad-4bdd-a920-390bc4701f0f-kube-api-access-zkgbp\") on node \"crc\" DevicePath \"\"" Feb 14 11:41:07 crc kubenswrapper[4904]: I0214 11:41:07.577407 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fa860db-2aad-4bdd-a920-390bc4701f0f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:41:07 crc kubenswrapper[4904]: I0214 11:41:07.577416 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fa860db-2aad-4bdd-a920-390bc4701f0f-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:41:07 crc kubenswrapper[4904]: I0214 11:41:07.942974 4904 generic.go:334] "Generic (PLEG): container finished" podID="1fa860db-2aad-4bdd-a920-390bc4701f0f" containerID="0bb4ad306413108c6d3ac874261aade422dfe0ee7216648679726d6272c0c32b" exitCode=0 Feb 14 11:41:07 crc kubenswrapper[4904]: I0214 11:41:07.943028 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzsr8" event={"ID":"1fa860db-2aad-4bdd-a920-390bc4701f0f","Type":"ContainerDied","Data":"0bb4ad306413108c6d3ac874261aade422dfe0ee7216648679726d6272c0c32b"} Feb 14 11:41:07 crc kubenswrapper[4904]: I0214 11:41:07.943098 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzsr8" event={"ID":"1fa860db-2aad-4bdd-a920-390bc4701f0f","Type":"ContainerDied","Data":"01dd16a9846350e47d454fde0cf4af6c7e23c3c6141d7909b95232d78f4f75b5"} Feb 14 11:41:07 crc kubenswrapper[4904]: I0214 11:41:07.943176 4904 scope.go:117] "RemoveContainer" containerID="0bb4ad306413108c6d3ac874261aade422dfe0ee7216648679726d6272c0c32b" Feb 14 11:41:07 crc kubenswrapper[4904]: I0214 11:41:07.944051 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mzsr8" Feb 14 11:41:07 crc kubenswrapper[4904]: I0214 11:41:07.966029 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzsr8"] Feb 14 11:41:07 crc kubenswrapper[4904]: I0214 11:41:07.974437 4904 scope.go:117] "RemoveContainer" containerID="eb536480ede3d5c2124f3c1899e875214c8235c2fe04933f95415cff5f2a7dc7" Feb 14 11:41:07 crc kubenswrapper[4904]: I0214 11:41:07.993208 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzsr8"] Feb 14 11:41:08 crc kubenswrapper[4904]: I0214 11:41:08.009309 4904 scope.go:117] "RemoveContainer" containerID="6e0e9ef14ca959e94259e0ffe0cff873e397ef03974a907d076f39b2c9a119f4" Feb 14 11:41:08 crc kubenswrapper[4904]: I0214 11:41:08.034559 4904 scope.go:117] "RemoveContainer" containerID="0bb4ad306413108c6d3ac874261aade422dfe0ee7216648679726d6272c0c32b" Feb 14 11:41:08 crc kubenswrapper[4904]: E0214 11:41:08.034997 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bb4ad306413108c6d3ac874261aade422dfe0ee7216648679726d6272c0c32b\": container with ID starting with 0bb4ad306413108c6d3ac874261aade422dfe0ee7216648679726d6272c0c32b not found: ID does not exist" containerID="0bb4ad306413108c6d3ac874261aade422dfe0ee7216648679726d6272c0c32b" Feb 14 11:41:08 crc kubenswrapper[4904]: I0214 11:41:08.035033 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bb4ad306413108c6d3ac874261aade422dfe0ee7216648679726d6272c0c32b"} err="failed to get container status \"0bb4ad306413108c6d3ac874261aade422dfe0ee7216648679726d6272c0c32b\": rpc error: code = NotFound desc = could not find container \"0bb4ad306413108c6d3ac874261aade422dfe0ee7216648679726d6272c0c32b\": container with ID starting with 0bb4ad306413108c6d3ac874261aade422dfe0ee7216648679726d6272c0c32b not found: ID does not exist" Feb 14 11:41:08 crc kubenswrapper[4904]: I0214 11:41:08.035054 4904 scope.go:117] "RemoveContainer" containerID="eb536480ede3d5c2124f3c1899e875214c8235c2fe04933f95415cff5f2a7dc7" Feb 14 11:41:08 crc kubenswrapper[4904]: E0214 11:41:08.035291 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb536480ede3d5c2124f3c1899e875214c8235c2fe04933f95415cff5f2a7dc7\": container with ID starting with eb536480ede3d5c2124f3c1899e875214c8235c2fe04933f95415cff5f2a7dc7 not found: ID does not exist" containerID="eb536480ede3d5c2124f3c1899e875214c8235c2fe04933f95415cff5f2a7dc7" Feb 14 11:41:08 crc kubenswrapper[4904]: I0214 11:41:08.035314 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb536480ede3d5c2124f3c1899e875214c8235c2fe04933f95415cff5f2a7dc7"} err="failed to get container status \"eb536480ede3d5c2124f3c1899e875214c8235c2fe04933f95415cff5f2a7dc7\": rpc error: code = NotFound desc = could not find container \"eb536480ede3d5c2124f3c1899e875214c8235c2fe04933f95415cff5f2a7dc7\": container with ID starting with eb536480ede3d5c2124f3c1899e875214c8235c2fe04933f95415cff5f2a7dc7 not found: ID does not exist" Feb 14 11:41:08 crc kubenswrapper[4904]: I0214 11:41:08.035327 4904 scope.go:117] "RemoveContainer" containerID="6e0e9ef14ca959e94259e0ffe0cff873e397ef03974a907d076f39b2c9a119f4" Feb 14 11:41:08 crc kubenswrapper[4904]: E0214 11:41:08.035518 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e0e9ef14ca959e94259e0ffe0cff873e397ef03974a907d076f39b2c9a119f4\": container with ID starting with 6e0e9ef14ca959e94259e0ffe0cff873e397ef03974a907d076f39b2c9a119f4 not found: ID does not exist" containerID="6e0e9ef14ca959e94259e0ffe0cff873e397ef03974a907d076f39b2c9a119f4" Feb 14 11:41:08 crc kubenswrapper[4904]: I0214 11:41:08.035535 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e0e9ef14ca959e94259e0ffe0cff873e397ef03974a907d076f39b2c9a119f4"} err="failed to get container status \"6e0e9ef14ca959e94259e0ffe0cff873e397ef03974a907d076f39b2c9a119f4\": rpc error: code = NotFound desc = could not find container \"6e0e9ef14ca959e94259e0ffe0cff873e397ef03974a907d076f39b2c9a119f4\": container with ID starting with 6e0e9ef14ca959e94259e0ffe0cff873e397ef03974a907d076f39b2c9a119f4 not found: ID does not exist" Feb 14 11:41:09 crc kubenswrapper[4904]: I0214 11:41:09.849378 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fa860db-2aad-4bdd-a920-390bc4701f0f" path="/var/lib/kubelet/pods/1fa860db-2aad-4bdd-a920-390bc4701f0f/volumes" Feb 14 11:41:54 crc kubenswrapper[4904]: I0214 11:41:54.380029 4904 generic.go:334] "Generic (PLEG): container finished" podID="b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6" containerID="ff6a1082a3b13f91f05388de46187529da853267e96357f44d90a2be875987fa" exitCode=2 Feb 14 11:41:54 crc kubenswrapper[4904]: I0214 11:41:54.380546 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" event={"ID":"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6","Type":"ContainerDied","Data":"ff6a1082a3b13f91f05388de46187529da853267e96357f44d90a2be875987fa"} Feb 14 11:41:55 crc kubenswrapper[4904]: I0214 11:41:55.763122 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" Feb 14 11:41:55 crc kubenswrapper[4904]: I0214 11:41:55.905597 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-repo-setup-combined-ca-bundle\") pod \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\" (UID: \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\") " Feb 14 11:41:55 crc kubenswrapper[4904]: I0214 11:41:55.905860 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmjxq\" (UniqueName: \"kubernetes.io/projected/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-kube-api-access-mmjxq\") pod \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\" (UID: \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\") " Feb 14 11:41:55 crc kubenswrapper[4904]: I0214 11:41:55.905946 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-ssh-key-openstack-edpm-ipam\") pod \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\" (UID: \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\") " Feb 14 11:41:55 crc kubenswrapper[4904]: I0214 11:41:55.906030 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-inventory\") pod \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\" (UID: \"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6\") " Feb 14 11:41:55 crc kubenswrapper[4904]: I0214 11:41:55.911246 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-kube-api-access-mmjxq" (OuterVolumeSpecName: "kube-api-access-mmjxq") pod "b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6" (UID: "b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6"). InnerVolumeSpecName "kube-api-access-mmjxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:41:55 crc kubenswrapper[4904]: I0214 11:41:55.914123 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6" (UID: "b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:41:55 crc kubenswrapper[4904]: I0214 11:41:55.936567 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-inventory" (OuterVolumeSpecName: "inventory") pod "b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6" (UID: "b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:41:55 crc kubenswrapper[4904]: I0214 11:41:55.937210 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6" (UID: "b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:41:56 crc kubenswrapper[4904]: I0214 11:41:56.008505 4904 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:41:56 crc kubenswrapper[4904]: I0214 11:41:56.008545 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmjxq\" (UniqueName: \"kubernetes.io/projected/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-kube-api-access-mmjxq\") on node \"crc\" DevicePath \"\"" Feb 14 11:41:56 crc kubenswrapper[4904]: I0214 11:41:56.008556 4904 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 14 11:41:56 crc kubenswrapper[4904]: I0214 11:41:56.008568 4904 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6-inventory\") on node \"crc\" DevicePath \"\"" Feb 14 11:41:56 crc kubenswrapper[4904]: I0214 11:41:56.394326 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" event={"ID":"b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6","Type":"ContainerDied","Data":"38ea72e8edea1d88df793a773936e43807490d960a8d13906cd530db7aafb897"} Feb 14 11:41:56 crc kubenswrapper[4904]: I0214 11:41:56.394618 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38ea72e8edea1d88df793a773936e43807490d960a8d13906cd530db7aafb897" Feb 14 11:41:56 crc kubenswrapper[4904]: I0214 11:41:56.394668 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.026239 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg"] Feb 14 11:43:13 crc kubenswrapper[4904]: E0214 11:43:13.027032 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.027046 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 11:43:13 crc kubenswrapper[4904]: E0214 11:43:13.027065 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a96f426c-8576-4f0b-8d83-f554dece8528" containerName="extract-content" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.027071 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="a96f426c-8576-4f0b-8d83-f554dece8528" containerName="extract-content" Feb 14 11:43:13 crc kubenswrapper[4904]: E0214 11:43:13.027085 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fa860db-2aad-4bdd-a920-390bc4701f0f" containerName="extract-utilities" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.027091 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fa860db-2aad-4bdd-a920-390bc4701f0f" containerName="extract-utilities" Feb 14 11:43:13 crc kubenswrapper[4904]: E0214 11:43:13.027102 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a96f426c-8576-4f0b-8d83-f554dece8528" containerName="extract-utilities" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.027107 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="a96f426c-8576-4f0b-8d83-f554dece8528" containerName="extract-utilities" Feb 14 11:43:13 crc kubenswrapper[4904]: E0214 11:43:13.027117 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a96f426c-8576-4f0b-8d83-f554dece8528" containerName="registry-server" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.027123 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="a96f426c-8576-4f0b-8d83-f554dece8528" containerName="registry-server" Feb 14 11:43:13 crc kubenswrapper[4904]: E0214 11:43:13.027132 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fa860db-2aad-4bdd-a920-390bc4701f0f" containerName="registry-server" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.027140 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fa860db-2aad-4bdd-a920-390bc4701f0f" containerName="registry-server" Feb 14 11:43:13 crc kubenswrapper[4904]: E0214 11:43:13.027189 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fa860db-2aad-4bdd-a920-390bc4701f0f" containerName="extract-content" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.027195 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fa860db-2aad-4bdd-a920-390bc4701f0f" containerName="extract-content" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.027360 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="a96f426c-8576-4f0b-8d83-f554dece8528" containerName="registry-server" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.027446 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.027459 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fa860db-2aad-4bdd-a920-390bc4701f0f" containerName="registry-server" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.028100 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.030082 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-48rnz" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.030887 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.031358 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.032580 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.048016 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg"] Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.098510 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d840dda5-8ddf-4efb-b121-e3113f0302ff-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg\" (UID: \"d840dda5-8ddf-4efb-b121-e3113f0302ff\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.098564 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbjf2\" (UniqueName: \"kubernetes.io/projected/d840dda5-8ddf-4efb-b121-e3113f0302ff-kube-api-access-gbjf2\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg\" (UID: \"d840dda5-8ddf-4efb-b121-e3113f0302ff\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.098607 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d840dda5-8ddf-4efb-b121-e3113f0302ff-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg\" (UID: \"d840dda5-8ddf-4efb-b121-e3113f0302ff\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.098632 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d840dda5-8ddf-4efb-b121-e3113f0302ff-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg\" (UID: \"d840dda5-8ddf-4efb-b121-e3113f0302ff\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.200596 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d840dda5-8ddf-4efb-b121-e3113f0302ff-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg\" (UID: \"d840dda5-8ddf-4efb-b121-e3113f0302ff\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.200728 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d840dda5-8ddf-4efb-b121-e3113f0302ff-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg\" (UID: \"d840dda5-8ddf-4efb-b121-e3113f0302ff\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.200764 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbjf2\" (UniqueName: \"kubernetes.io/projected/d840dda5-8ddf-4efb-b121-e3113f0302ff-kube-api-access-gbjf2\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg\" (UID: \"d840dda5-8ddf-4efb-b121-e3113f0302ff\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.200809 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d840dda5-8ddf-4efb-b121-e3113f0302ff-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg\" (UID: \"d840dda5-8ddf-4efb-b121-e3113f0302ff\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.211179 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d840dda5-8ddf-4efb-b121-e3113f0302ff-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg\" (UID: \"d840dda5-8ddf-4efb-b121-e3113f0302ff\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.214808 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d840dda5-8ddf-4efb-b121-e3113f0302ff-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg\" (UID: \"d840dda5-8ddf-4efb-b121-e3113f0302ff\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.216796 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d840dda5-8ddf-4efb-b121-e3113f0302ff-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg\" (UID: \"d840dda5-8ddf-4efb-b121-e3113f0302ff\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.222515 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbjf2\" (UniqueName: \"kubernetes.io/projected/d840dda5-8ddf-4efb-b121-e3113f0302ff-kube-api-access-gbjf2\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg\" (UID: \"d840dda5-8ddf-4efb-b121-e3113f0302ff\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.345141 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.847175 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg"] Feb 14 11:43:13 crc kubenswrapper[4904]: I0214 11:43:13.874747 4904 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 14 11:43:14 crc kubenswrapper[4904]: I0214 11:43:14.003917 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" event={"ID":"d840dda5-8ddf-4efb-b121-e3113f0302ff","Type":"ContainerStarted","Data":"cdb128c57d2d88358bd7945ea3120569f5c983219efc76950903ba5ae78ae8fe"} Feb 14 11:43:15 crc kubenswrapper[4904]: I0214 11:43:15.013054 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" event={"ID":"d840dda5-8ddf-4efb-b121-e3113f0302ff","Type":"ContainerStarted","Data":"26d7877877595b77d4cba9db68011d50f0a2e8225a74278ec4da4c4ac998a4d4"} Feb 14 11:43:15 crc kubenswrapper[4904]: I0214 11:43:15.032878 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" podStartSLOduration=1.540060842 podStartE2EDuration="2.032856374s" podCreationTimestamp="2026-02-14 11:43:13 +0000 UTC" firstStartedPulling="2026-02-14 11:43:13.874537236 +0000 UTC m=+1984.687301897" lastFinishedPulling="2026-02-14 11:43:14.367332738 +0000 UTC m=+1985.180097429" observedRunningTime="2026-02-14 11:43:15.030436099 +0000 UTC m=+1985.843200760" watchObservedRunningTime="2026-02-14 11:43:15.032856374 +0000 UTC m=+1985.845621055" Feb 14 11:43:16 crc kubenswrapper[4904]: I0214 11:43:16.382866 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:43:16 crc kubenswrapper[4904]: I0214 11:43:16.383231 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:43:46 crc kubenswrapper[4904]: I0214 11:43:46.382730 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:43:46 crc kubenswrapper[4904]: I0214 11:43:46.383317 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:44:16 crc kubenswrapper[4904]: I0214 11:44:16.382439 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:44:16 crc kubenswrapper[4904]: I0214 11:44:16.383052 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:44:16 crc kubenswrapper[4904]: I0214 11:44:16.383137 4904 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:44:16 crc kubenswrapper[4904]: I0214 11:44:16.383641 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"36065fedbbfd5007e91cc94e16b80606571b96f7d01d643f9259259819fe5f66"} pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 14 11:44:16 crc kubenswrapper[4904]: I0214 11:44:16.383696 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" containerID="cri-o://36065fedbbfd5007e91cc94e16b80606571b96f7d01d643f9259259819fe5f66" gracePeriod=600 Feb 14 11:44:16 crc kubenswrapper[4904]: I0214 11:44:16.510148 4904 generic.go:334] "Generic (PLEG): container finished" podID="7790d276-2506-4641-8c22-67e6441417f0" containerID="36065fedbbfd5007e91cc94e16b80606571b96f7d01d643f9259259819fe5f66" exitCode=0 Feb 14 11:44:16 crc kubenswrapper[4904]: I0214 11:44:16.510213 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerDied","Data":"36065fedbbfd5007e91cc94e16b80606571b96f7d01d643f9259259819fe5f66"} Feb 14 11:44:16 crc kubenswrapper[4904]: I0214 11:44:16.510258 4904 scope.go:117] "RemoveContainer" containerID="6fee523332ea2722edd2eda02be775290583b333d2a71ec990d83d77c4fb86fc" Feb 14 11:44:17 crc kubenswrapper[4904]: I0214 11:44:17.520016 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb"} Feb 14 11:45:00 crc kubenswrapper[4904]: I0214 11:45:00.146907 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l"] Feb 14 11:45:00 crc kubenswrapper[4904]: I0214 11:45:00.148612 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l" Feb 14 11:45:00 crc kubenswrapper[4904]: I0214 11:45:00.154032 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 14 11:45:00 crc kubenswrapper[4904]: I0214 11:45:00.159171 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l"] Feb 14 11:45:00 crc kubenswrapper[4904]: I0214 11:45:00.160491 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 14 11:45:00 crc kubenswrapper[4904]: I0214 11:45:00.202373 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c62364cf-6155-4246-be2f-48bb500d9fb7-config-volume\") pod \"collect-profiles-29517825-wzv4l\" (UID: \"c62364cf-6155-4246-be2f-48bb500d9fb7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l" Feb 14 11:45:00 crc kubenswrapper[4904]: I0214 11:45:00.202435 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c62364cf-6155-4246-be2f-48bb500d9fb7-secret-volume\") pod \"collect-profiles-29517825-wzv4l\" (UID: \"c62364cf-6155-4246-be2f-48bb500d9fb7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l" Feb 14 11:45:00 crc kubenswrapper[4904]: I0214 11:45:00.202697 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r88q6\" (UniqueName: \"kubernetes.io/projected/c62364cf-6155-4246-be2f-48bb500d9fb7-kube-api-access-r88q6\") pod \"collect-profiles-29517825-wzv4l\" (UID: \"c62364cf-6155-4246-be2f-48bb500d9fb7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l" Feb 14 11:45:00 crc kubenswrapper[4904]: I0214 11:45:00.304909 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c62364cf-6155-4246-be2f-48bb500d9fb7-config-volume\") pod \"collect-profiles-29517825-wzv4l\" (UID: \"c62364cf-6155-4246-be2f-48bb500d9fb7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l" Feb 14 11:45:00 crc kubenswrapper[4904]: I0214 11:45:00.304960 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c62364cf-6155-4246-be2f-48bb500d9fb7-secret-volume\") pod \"collect-profiles-29517825-wzv4l\" (UID: \"c62364cf-6155-4246-be2f-48bb500d9fb7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l" Feb 14 11:45:00 crc kubenswrapper[4904]: I0214 11:45:00.305046 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r88q6\" (UniqueName: \"kubernetes.io/projected/c62364cf-6155-4246-be2f-48bb500d9fb7-kube-api-access-r88q6\") pod \"collect-profiles-29517825-wzv4l\" (UID: \"c62364cf-6155-4246-be2f-48bb500d9fb7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l" Feb 14 11:45:00 crc kubenswrapper[4904]: I0214 11:45:00.305730 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c62364cf-6155-4246-be2f-48bb500d9fb7-config-volume\") pod \"collect-profiles-29517825-wzv4l\" (UID: \"c62364cf-6155-4246-be2f-48bb500d9fb7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l" Feb 14 11:45:00 crc kubenswrapper[4904]: I0214 11:45:00.318081 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c62364cf-6155-4246-be2f-48bb500d9fb7-secret-volume\") pod \"collect-profiles-29517825-wzv4l\" (UID: \"c62364cf-6155-4246-be2f-48bb500d9fb7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l" Feb 14 11:45:00 crc kubenswrapper[4904]: I0214 11:45:00.325409 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r88q6\" (UniqueName: \"kubernetes.io/projected/c62364cf-6155-4246-be2f-48bb500d9fb7-kube-api-access-r88q6\") pod \"collect-profiles-29517825-wzv4l\" (UID: \"c62364cf-6155-4246-be2f-48bb500d9fb7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l" Feb 14 11:45:00 crc kubenswrapper[4904]: I0214 11:45:00.477598 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l" Feb 14 11:45:00 crc kubenswrapper[4904]: I0214 11:45:00.941593 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l"] Feb 14 11:45:01 crc kubenswrapper[4904]: I0214 11:45:01.871418 4904 generic.go:334] "Generic (PLEG): container finished" podID="c62364cf-6155-4246-be2f-48bb500d9fb7" containerID="5733dc3b95e94b18e6c617fe41c7000f51016214880747bc947afa1962e680b8" exitCode=0 Feb 14 11:45:01 crc kubenswrapper[4904]: I0214 11:45:01.871469 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l" event={"ID":"c62364cf-6155-4246-be2f-48bb500d9fb7","Type":"ContainerDied","Data":"5733dc3b95e94b18e6c617fe41c7000f51016214880747bc947afa1962e680b8"} Feb 14 11:45:01 crc kubenswrapper[4904]: I0214 11:45:01.871718 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l" event={"ID":"c62364cf-6155-4246-be2f-48bb500d9fb7","Type":"ContainerStarted","Data":"0ebe239f4d838d95513fdef7c156857f54a1d88d39a5dc3860f928f3db66958a"} Feb 14 11:45:03 crc kubenswrapper[4904]: I0214 11:45:03.251707 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l" Feb 14 11:45:03 crc kubenswrapper[4904]: I0214 11:45:03.360962 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c62364cf-6155-4246-be2f-48bb500d9fb7-secret-volume\") pod \"c62364cf-6155-4246-be2f-48bb500d9fb7\" (UID: \"c62364cf-6155-4246-be2f-48bb500d9fb7\") " Feb 14 11:45:03 crc kubenswrapper[4904]: I0214 11:45:03.361246 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r88q6\" (UniqueName: \"kubernetes.io/projected/c62364cf-6155-4246-be2f-48bb500d9fb7-kube-api-access-r88q6\") pod \"c62364cf-6155-4246-be2f-48bb500d9fb7\" (UID: \"c62364cf-6155-4246-be2f-48bb500d9fb7\") " Feb 14 11:45:03 crc kubenswrapper[4904]: I0214 11:45:03.361306 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c62364cf-6155-4246-be2f-48bb500d9fb7-config-volume\") pod \"c62364cf-6155-4246-be2f-48bb500d9fb7\" (UID: \"c62364cf-6155-4246-be2f-48bb500d9fb7\") " Feb 14 11:45:03 crc kubenswrapper[4904]: I0214 11:45:03.362422 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c62364cf-6155-4246-be2f-48bb500d9fb7-config-volume" (OuterVolumeSpecName: "config-volume") pod "c62364cf-6155-4246-be2f-48bb500d9fb7" (UID: "c62364cf-6155-4246-be2f-48bb500d9fb7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 11:45:03 crc kubenswrapper[4904]: I0214 11:45:03.368977 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c62364cf-6155-4246-be2f-48bb500d9fb7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c62364cf-6155-4246-be2f-48bb500d9fb7" (UID: "c62364cf-6155-4246-be2f-48bb500d9fb7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:45:03 crc kubenswrapper[4904]: I0214 11:45:03.370098 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c62364cf-6155-4246-be2f-48bb500d9fb7-kube-api-access-r88q6" (OuterVolumeSpecName: "kube-api-access-r88q6") pod "c62364cf-6155-4246-be2f-48bb500d9fb7" (UID: "c62364cf-6155-4246-be2f-48bb500d9fb7"). InnerVolumeSpecName "kube-api-access-r88q6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:45:03 crc kubenswrapper[4904]: I0214 11:45:03.463696 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r88q6\" (UniqueName: \"kubernetes.io/projected/c62364cf-6155-4246-be2f-48bb500d9fb7-kube-api-access-r88q6\") on node \"crc\" DevicePath \"\"" Feb 14 11:45:03 crc kubenswrapper[4904]: I0214 11:45:03.463740 4904 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c62364cf-6155-4246-be2f-48bb500d9fb7-config-volume\") on node \"crc\" DevicePath \"\"" Feb 14 11:45:03 crc kubenswrapper[4904]: I0214 11:45:03.463752 4904 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c62364cf-6155-4246-be2f-48bb500d9fb7-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 14 11:45:03 crc kubenswrapper[4904]: I0214 11:45:03.887972 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l" event={"ID":"c62364cf-6155-4246-be2f-48bb500d9fb7","Type":"ContainerDied","Data":"0ebe239f4d838d95513fdef7c156857f54a1d88d39a5dc3860f928f3db66958a"} Feb 14 11:45:03 crc kubenswrapper[4904]: I0214 11:45:03.888317 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ebe239f4d838d95513fdef7c156857f54a1d88d39a5dc3860f928f3db66958a" Feb 14 11:45:03 crc kubenswrapper[4904]: I0214 11:45:03.888207 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l" Feb 14 11:45:04 crc kubenswrapper[4904]: I0214 11:45:04.363648 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx"] Feb 14 11:45:04 crc kubenswrapper[4904]: I0214 11:45:04.373478 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517780-v65bx"] Feb 14 11:45:05 crc kubenswrapper[4904]: I0214 11:45:05.849425 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d9f5d3d-fb18-4302-8b6b-62d4b8999b65" path="/var/lib/kubelet/pods/6d9f5d3d-fb18-4302-8b6b-62d4b8999b65/volumes" Feb 14 11:45:15 crc kubenswrapper[4904]: I0214 11:45:15.803677 4904 scope.go:117] "RemoveContainer" containerID="6a6ed2121e9a25309d4d0411dd1abe47ea2f42a715eae146c16d359ff323786c" Feb 14 11:45:39 crc kubenswrapper[4904]: I0214 11:45:39.236744 4904 generic.go:334] "Generic (PLEG): container finished" podID="d840dda5-8ddf-4efb-b121-e3113f0302ff" containerID="26d7877877595b77d4cba9db68011d50f0a2e8225a74278ec4da4c4ac998a4d4" exitCode=2 Feb 14 11:45:39 crc kubenswrapper[4904]: I0214 11:45:39.236873 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" event={"ID":"d840dda5-8ddf-4efb-b121-e3113f0302ff","Type":"ContainerDied","Data":"26d7877877595b77d4cba9db68011d50f0a2e8225a74278ec4da4c4ac998a4d4"} Feb 14 11:45:40 crc kubenswrapper[4904]: I0214 11:45:40.677516 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" Feb 14 11:45:40 crc kubenswrapper[4904]: I0214 11:45:40.778263 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d840dda5-8ddf-4efb-b121-e3113f0302ff-inventory\") pod \"d840dda5-8ddf-4efb-b121-e3113f0302ff\" (UID: \"d840dda5-8ddf-4efb-b121-e3113f0302ff\") " Feb 14 11:45:40 crc kubenswrapper[4904]: I0214 11:45:40.778617 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbjf2\" (UniqueName: \"kubernetes.io/projected/d840dda5-8ddf-4efb-b121-e3113f0302ff-kube-api-access-gbjf2\") pod \"d840dda5-8ddf-4efb-b121-e3113f0302ff\" (UID: \"d840dda5-8ddf-4efb-b121-e3113f0302ff\") " Feb 14 11:45:40 crc kubenswrapper[4904]: I0214 11:45:40.778711 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d840dda5-8ddf-4efb-b121-e3113f0302ff-ssh-key-openstack-edpm-ipam\") pod \"d840dda5-8ddf-4efb-b121-e3113f0302ff\" (UID: \"d840dda5-8ddf-4efb-b121-e3113f0302ff\") " Feb 14 11:45:40 crc kubenswrapper[4904]: I0214 11:45:40.778798 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d840dda5-8ddf-4efb-b121-e3113f0302ff-repo-setup-combined-ca-bundle\") pod \"d840dda5-8ddf-4efb-b121-e3113f0302ff\" (UID: \"d840dda5-8ddf-4efb-b121-e3113f0302ff\") " Feb 14 11:45:40 crc kubenswrapper[4904]: I0214 11:45:40.801369 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d840dda5-8ddf-4efb-b121-e3113f0302ff-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "d840dda5-8ddf-4efb-b121-e3113f0302ff" (UID: "d840dda5-8ddf-4efb-b121-e3113f0302ff"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:45:40 crc kubenswrapper[4904]: I0214 11:45:40.808608 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d840dda5-8ddf-4efb-b121-e3113f0302ff-kube-api-access-gbjf2" (OuterVolumeSpecName: "kube-api-access-gbjf2") pod "d840dda5-8ddf-4efb-b121-e3113f0302ff" (UID: "d840dda5-8ddf-4efb-b121-e3113f0302ff"). InnerVolumeSpecName "kube-api-access-gbjf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:45:40 crc kubenswrapper[4904]: I0214 11:45:40.811657 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d840dda5-8ddf-4efb-b121-e3113f0302ff-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d840dda5-8ddf-4efb-b121-e3113f0302ff" (UID: "d840dda5-8ddf-4efb-b121-e3113f0302ff"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:45:40 crc kubenswrapper[4904]: I0214 11:45:40.813573 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d840dda5-8ddf-4efb-b121-e3113f0302ff-inventory" (OuterVolumeSpecName: "inventory") pod "d840dda5-8ddf-4efb-b121-e3113f0302ff" (UID: "d840dda5-8ddf-4efb-b121-e3113f0302ff"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:45:40 crc kubenswrapper[4904]: I0214 11:45:40.880886 4904 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d840dda5-8ddf-4efb-b121-e3113f0302ff-inventory\") on node \"crc\" DevicePath \"\"" Feb 14 11:45:40 crc kubenswrapper[4904]: I0214 11:45:40.880914 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbjf2\" (UniqueName: \"kubernetes.io/projected/d840dda5-8ddf-4efb-b121-e3113f0302ff-kube-api-access-gbjf2\") on node \"crc\" DevicePath \"\"" Feb 14 11:45:40 crc kubenswrapper[4904]: I0214 11:45:40.880928 4904 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d840dda5-8ddf-4efb-b121-e3113f0302ff-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 14 11:45:40 crc kubenswrapper[4904]: I0214 11:45:40.880940 4904 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d840dda5-8ddf-4efb-b121-e3113f0302ff-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:45:41 crc kubenswrapper[4904]: I0214 11:45:41.252546 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" event={"ID":"d840dda5-8ddf-4efb-b121-e3113f0302ff","Type":"ContainerDied","Data":"cdb128c57d2d88358bd7945ea3120569f5c983219efc76950903ba5ae78ae8fe"} Feb 14 11:45:41 crc kubenswrapper[4904]: I0214 11:45:41.252582 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cdb128c57d2d88358bd7945ea3120569f5c983219efc76950903ba5ae78ae8fe" Feb 14 11:45:41 crc kubenswrapper[4904]: I0214 11:45:41.252623 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg" Feb 14 11:46:16 crc kubenswrapper[4904]: I0214 11:46:16.382521 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:46:16 crc kubenswrapper[4904]: I0214 11:46:16.383143 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:46:46 crc kubenswrapper[4904]: I0214 11:46:46.383134 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:46:46 crc kubenswrapper[4904]: I0214 11:46:46.383527 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:47:02 crc kubenswrapper[4904]: I0214 11:47:02.029296 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p4czx"] Feb 14 11:47:02 crc kubenswrapper[4904]: E0214 11:47:02.030375 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c62364cf-6155-4246-be2f-48bb500d9fb7" containerName="collect-profiles" Feb 14 11:47:02 crc kubenswrapper[4904]: I0214 11:47:02.030392 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="c62364cf-6155-4246-be2f-48bb500d9fb7" containerName="collect-profiles" Feb 14 11:47:02 crc kubenswrapper[4904]: E0214 11:47:02.030414 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d840dda5-8ddf-4efb-b121-e3113f0302ff" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 11:47:02 crc kubenswrapper[4904]: I0214 11:47:02.030424 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="d840dda5-8ddf-4efb-b121-e3113f0302ff" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 11:47:02 crc kubenswrapper[4904]: I0214 11:47:02.030690 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="c62364cf-6155-4246-be2f-48bb500d9fb7" containerName="collect-profiles" Feb 14 11:47:02 crc kubenswrapper[4904]: I0214 11:47:02.030725 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="d840dda5-8ddf-4efb-b121-e3113f0302ff" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 11:47:02 crc kubenswrapper[4904]: I0214 11:47:02.032778 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p4czx" Feb 14 11:47:02 crc kubenswrapper[4904]: I0214 11:47:02.042672 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p4czx"] Feb 14 11:47:02 crc kubenswrapper[4904]: I0214 11:47:02.188588 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnmng\" (UniqueName: \"kubernetes.io/projected/13524bff-0982-4b53-9f63-f492a596236e-kube-api-access-gnmng\") pod \"redhat-operators-p4czx\" (UID: \"13524bff-0982-4b53-9f63-f492a596236e\") " pod="openshift-marketplace/redhat-operators-p4czx" Feb 14 11:47:02 crc kubenswrapper[4904]: I0214 11:47:02.188969 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13524bff-0982-4b53-9f63-f492a596236e-utilities\") pod \"redhat-operators-p4czx\" (UID: \"13524bff-0982-4b53-9f63-f492a596236e\") " pod="openshift-marketplace/redhat-operators-p4czx" Feb 14 11:47:02 crc kubenswrapper[4904]: I0214 11:47:02.189114 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13524bff-0982-4b53-9f63-f492a596236e-catalog-content\") pod \"redhat-operators-p4czx\" (UID: \"13524bff-0982-4b53-9f63-f492a596236e\") " pod="openshift-marketplace/redhat-operators-p4czx" Feb 14 11:47:02 crc kubenswrapper[4904]: I0214 11:47:02.291373 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnmng\" (UniqueName: \"kubernetes.io/projected/13524bff-0982-4b53-9f63-f492a596236e-kube-api-access-gnmng\") pod \"redhat-operators-p4czx\" (UID: \"13524bff-0982-4b53-9f63-f492a596236e\") " pod="openshift-marketplace/redhat-operators-p4czx" Feb 14 11:47:02 crc kubenswrapper[4904]: I0214 11:47:02.291501 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13524bff-0982-4b53-9f63-f492a596236e-utilities\") pod \"redhat-operators-p4czx\" (UID: \"13524bff-0982-4b53-9f63-f492a596236e\") " pod="openshift-marketplace/redhat-operators-p4czx" Feb 14 11:47:02 crc kubenswrapper[4904]: I0214 11:47:02.291559 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13524bff-0982-4b53-9f63-f492a596236e-catalog-content\") pod \"redhat-operators-p4czx\" (UID: \"13524bff-0982-4b53-9f63-f492a596236e\") " pod="openshift-marketplace/redhat-operators-p4czx" Feb 14 11:47:02 crc kubenswrapper[4904]: I0214 11:47:02.292697 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13524bff-0982-4b53-9f63-f492a596236e-catalog-content\") pod \"redhat-operators-p4czx\" (UID: \"13524bff-0982-4b53-9f63-f492a596236e\") " pod="openshift-marketplace/redhat-operators-p4czx" Feb 14 11:47:02 crc kubenswrapper[4904]: I0214 11:47:02.292849 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13524bff-0982-4b53-9f63-f492a596236e-utilities\") pod \"redhat-operators-p4czx\" (UID: \"13524bff-0982-4b53-9f63-f492a596236e\") " pod="openshift-marketplace/redhat-operators-p4czx" Feb 14 11:47:02 crc kubenswrapper[4904]: I0214 11:47:02.326437 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnmng\" (UniqueName: \"kubernetes.io/projected/13524bff-0982-4b53-9f63-f492a596236e-kube-api-access-gnmng\") pod \"redhat-operators-p4czx\" (UID: \"13524bff-0982-4b53-9f63-f492a596236e\") " pod="openshift-marketplace/redhat-operators-p4czx" Feb 14 11:47:02 crc kubenswrapper[4904]: I0214 11:47:02.402484 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p4czx" Feb 14 11:47:02 crc kubenswrapper[4904]: I0214 11:47:02.853758 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p4czx"] Feb 14 11:47:02 crc kubenswrapper[4904]: I0214 11:47:02.976498 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4czx" event={"ID":"13524bff-0982-4b53-9f63-f492a596236e","Type":"ContainerStarted","Data":"aae9a447ebfd1e4547bbe8b1e31306366eee012b5a71d2cfc5af418f9816e8ff"} Feb 14 11:47:03 crc kubenswrapper[4904]: I0214 11:47:03.987415 4904 generic.go:334] "Generic (PLEG): container finished" podID="13524bff-0982-4b53-9f63-f492a596236e" containerID="d0d46441c057b598c68f94a0a0009cf756f1760675151e8ca435ef704a57577a" exitCode=0 Feb 14 11:47:03 crc kubenswrapper[4904]: I0214 11:47:03.987574 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4czx" event={"ID":"13524bff-0982-4b53-9f63-f492a596236e","Type":"ContainerDied","Data":"d0d46441c057b598c68f94a0a0009cf756f1760675151e8ca435ef704a57577a"} Feb 14 11:47:05 crc kubenswrapper[4904]: I0214 11:47:05.005375 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4czx" event={"ID":"13524bff-0982-4b53-9f63-f492a596236e","Type":"ContainerStarted","Data":"f3aaa27d67574612846901271d2b3f13a1f48ee36c1a0d9cc327eefb9ce7de57"} Feb 14 11:47:09 crc kubenswrapper[4904]: I0214 11:47:09.043851 4904 generic.go:334] "Generic (PLEG): container finished" podID="13524bff-0982-4b53-9f63-f492a596236e" containerID="f3aaa27d67574612846901271d2b3f13a1f48ee36c1a0d9cc327eefb9ce7de57" exitCode=0 Feb 14 11:47:09 crc kubenswrapper[4904]: I0214 11:47:09.044077 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4czx" event={"ID":"13524bff-0982-4b53-9f63-f492a596236e","Type":"ContainerDied","Data":"f3aaa27d67574612846901271d2b3f13a1f48ee36c1a0d9cc327eefb9ce7de57"} Feb 14 11:47:10 crc kubenswrapper[4904]: I0214 11:47:10.071601 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4czx" event={"ID":"13524bff-0982-4b53-9f63-f492a596236e","Type":"ContainerStarted","Data":"dd2253a2b86148ff596cc012bcdb92a5f9890126d945011fbd980bdf2a86a1a3"} Feb 14 11:47:10 crc kubenswrapper[4904]: I0214 11:47:10.127460 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p4czx" podStartSLOduration=2.666977936 podStartE2EDuration="8.12743918s" podCreationTimestamp="2026-02-14 11:47:02 +0000 UTC" firstStartedPulling="2026-02-14 11:47:03.990258622 +0000 UTC m=+2214.803023293" lastFinishedPulling="2026-02-14 11:47:09.450719876 +0000 UTC m=+2220.263484537" observedRunningTime="2026-02-14 11:47:10.121936309 +0000 UTC m=+2220.934700970" watchObservedRunningTime="2026-02-14 11:47:10.12743918 +0000 UTC m=+2220.940203841" Feb 14 11:47:12 crc kubenswrapper[4904]: I0214 11:47:12.402981 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p4czx" Feb 14 11:47:12 crc kubenswrapper[4904]: I0214 11:47:12.403594 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p4czx" Feb 14 11:47:13 crc kubenswrapper[4904]: I0214 11:47:13.453051 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-p4czx" podUID="13524bff-0982-4b53-9f63-f492a596236e" containerName="registry-server" probeResult="failure" output=< Feb 14 11:47:13 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 11:47:13 crc kubenswrapper[4904]: > Feb 14 11:47:16 crc kubenswrapper[4904]: I0214 11:47:16.383283 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:47:16 crc kubenswrapper[4904]: I0214 11:47:16.383904 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:47:16 crc kubenswrapper[4904]: I0214 11:47:16.383960 4904 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:47:16 crc kubenswrapper[4904]: I0214 11:47:16.384969 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb"} pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 14 11:47:16 crc kubenswrapper[4904]: I0214 11:47:16.385046 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" containerID="cri-o://2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" gracePeriod=600 Feb 14 11:47:16 crc kubenswrapper[4904]: E0214 11:47:16.512684 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:47:17 crc kubenswrapper[4904]: I0214 11:47:17.127013 4904 generic.go:334] "Generic (PLEG): container finished" podID="7790d276-2506-4641-8c22-67e6441417f0" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" exitCode=0 Feb 14 11:47:17 crc kubenswrapper[4904]: I0214 11:47:17.127060 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerDied","Data":"2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb"} Feb 14 11:47:17 crc kubenswrapper[4904]: I0214 11:47:17.127100 4904 scope.go:117] "RemoveContainer" containerID="36065fedbbfd5007e91cc94e16b80606571b96f7d01d643f9259259819fe5f66" Feb 14 11:47:17 crc kubenswrapper[4904]: I0214 11:47:17.127633 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:47:17 crc kubenswrapper[4904]: E0214 11:47:17.127991 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:47:22 crc kubenswrapper[4904]: I0214 11:47:22.453199 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p4czx" Feb 14 11:47:22 crc kubenswrapper[4904]: I0214 11:47:22.510489 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p4czx" Feb 14 11:47:22 crc kubenswrapper[4904]: I0214 11:47:22.833566 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p4czx"] Feb 14 11:47:24 crc kubenswrapper[4904]: I0214 11:47:24.182660 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p4czx" podUID="13524bff-0982-4b53-9f63-f492a596236e" containerName="registry-server" containerID="cri-o://dd2253a2b86148ff596cc012bcdb92a5f9890126d945011fbd980bdf2a86a1a3" gracePeriod=2 Feb 14 11:47:24 crc kubenswrapper[4904]: I0214 11:47:24.724319 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p4czx" Feb 14 11:47:24 crc kubenswrapper[4904]: I0214 11:47:24.820262 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13524bff-0982-4b53-9f63-f492a596236e-catalog-content\") pod \"13524bff-0982-4b53-9f63-f492a596236e\" (UID: \"13524bff-0982-4b53-9f63-f492a596236e\") " Feb 14 11:47:24 crc kubenswrapper[4904]: I0214 11:47:24.820458 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13524bff-0982-4b53-9f63-f492a596236e-utilities\") pod \"13524bff-0982-4b53-9f63-f492a596236e\" (UID: \"13524bff-0982-4b53-9f63-f492a596236e\") " Feb 14 11:47:24 crc kubenswrapper[4904]: I0214 11:47:24.820518 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnmng\" (UniqueName: \"kubernetes.io/projected/13524bff-0982-4b53-9f63-f492a596236e-kube-api-access-gnmng\") pod \"13524bff-0982-4b53-9f63-f492a596236e\" (UID: \"13524bff-0982-4b53-9f63-f492a596236e\") " Feb 14 11:47:24 crc kubenswrapper[4904]: I0214 11:47:24.821416 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13524bff-0982-4b53-9f63-f492a596236e-utilities" (OuterVolumeSpecName: "utilities") pod "13524bff-0982-4b53-9f63-f492a596236e" (UID: "13524bff-0982-4b53-9f63-f492a596236e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:47:24 crc kubenswrapper[4904]: I0214 11:47:24.825956 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13524bff-0982-4b53-9f63-f492a596236e-kube-api-access-gnmng" (OuterVolumeSpecName: "kube-api-access-gnmng") pod "13524bff-0982-4b53-9f63-f492a596236e" (UID: "13524bff-0982-4b53-9f63-f492a596236e"). InnerVolumeSpecName "kube-api-access-gnmng". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:47:24 crc kubenswrapper[4904]: I0214 11:47:24.923216 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnmng\" (UniqueName: \"kubernetes.io/projected/13524bff-0982-4b53-9f63-f492a596236e-kube-api-access-gnmng\") on node \"crc\" DevicePath \"\"" Feb 14 11:47:24 crc kubenswrapper[4904]: I0214 11:47:24.923355 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13524bff-0982-4b53-9f63-f492a596236e-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:47:24 crc kubenswrapper[4904]: I0214 11:47:24.981746 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13524bff-0982-4b53-9f63-f492a596236e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "13524bff-0982-4b53-9f63-f492a596236e" (UID: "13524bff-0982-4b53-9f63-f492a596236e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:47:25 crc kubenswrapper[4904]: I0214 11:47:25.024751 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13524bff-0982-4b53-9f63-f492a596236e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:47:25 crc kubenswrapper[4904]: I0214 11:47:25.193799 4904 generic.go:334] "Generic (PLEG): container finished" podID="13524bff-0982-4b53-9f63-f492a596236e" containerID="dd2253a2b86148ff596cc012bcdb92a5f9890126d945011fbd980bdf2a86a1a3" exitCode=0 Feb 14 11:47:25 crc kubenswrapper[4904]: I0214 11:47:25.193859 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4czx" event={"ID":"13524bff-0982-4b53-9f63-f492a596236e","Type":"ContainerDied","Data":"dd2253a2b86148ff596cc012bcdb92a5f9890126d945011fbd980bdf2a86a1a3"} Feb 14 11:47:25 crc kubenswrapper[4904]: I0214 11:47:25.193892 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4czx" event={"ID":"13524bff-0982-4b53-9f63-f492a596236e","Type":"ContainerDied","Data":"aae9a447ebfd1e4547bbe8b1e31306366eee012b5a71d2cfc5af418f9816e8ff"} Feb 14 11:47:25 crc kubenswrapper[4904]: I0214 11:47:25.193933 4904 scope.go:117] "RemoveContainer" containerID="dd2253a2b86148ff596cc012bcdb92a5f9890126d945011fbd980bdf2a86a1a3" Feb 14 11:47:25 crc kubenswrapper[4904]: I0214 11:47:25.195108 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p4czx" Feb 14 11:47:25 crc kubenswrapper[4904]: I0214 11:47:25.237605 4904 scope.go:117] "RemoveContainer" containerID="f3aaa27d67574612846901271d2b3f13a1f48ee36c1a0d9cc327eefb9ce7de57" Feb 14 11:47:25 crc kubenswrapper[4904]: I0214 11:47:25.238312 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p4czx"] Feb 14 11:47:25 crc kubenswrapper[4904]: I0214 11:47:25.253509 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p4czx"] Feb 14 11:47:25 crc kubenswrapper[4904]: I0214 11:47:25.265079 4904 scope.go:117] "RemoveContainer" containerID="d0d46441c057b598c68f94a0a0009cf756f1760675151e8ca435ef704a57577a" Feb 14 11:47:25 crc kubenswrapper[4904]: I0214 11:47:25.309852 4904 scope.go:117] "RemoveContainer" containerID="dd2253a2b86148ff596cc012bcdb92a5f9890126d945011fbd980bdf2a86a1a3" Feb 14 11:47:25 crc kubenswrapper[4904]: E0214 11:47:25.310343 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd2253a2b86148ff596cc012bcdb92a5f9890126d945011fbd980bdf2a86a1a3\": container with ID starting with dd2253a2b86148ff596cc012bcdb92a5f9890126d945011fbd980bdf2a86a1a3 not found: ID does not exist" containerID="dd2253a2b86148ff596cc012bcdb92a5f9890126d945011fbd980bdf2a86a1a3" Feb 14 11:47:25 crc kubenswrapper[4904]: I0214 11:47:25.310381 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd2253a2b86148ff596cc012bcdb92a5f9890126d945011fbd980bdf2a86a1a3"} err="failed to get container status \"dd2253a2b86148ff596cc012bcdb92a5f9890126d945011fbd980bdf2a86a1a3\": rpc error: code = NotFound desc = could not find container \"dd2253a2b86148ff596cc012bcdb92a5f9890126d945011fbd980bdf2a86a1a3\": container with ID starting with dd2253a2b86148ff596cc012bcdb92a5f9890126d945011fbd980bdf2a86a1a3 not found: ID does not exist" Feb 14 11:47:25 crc kubenswrapper[4904]: I0214 11:47:25.310409 4904 scope.go:117] "RemoveContainer" containerID="f3aaa27d67574612846901271d2b3f13a1f48ee36c1a0d9cc327eefb9ce7de57" Feb 14 11:47:25 crc kubenswrapper[4904]: E0214 11:47:25.310696 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3aaa27d67574612846901271d2b3f13a1f48ee36c1a0d9cc327eefb9ce7de57\": container with ID starting with f3aaa27d67574612846901271d2b3f13a1f48ee36c1a0d9cc327eefb9ce7de57 not found: ID does not exist" containerID="f3aaa27d67574612846901271d2b3f13a1f48ee36c1a0d9cc327eefb9ce7de57" Feb 14 11:47:25 crc kubenswrapper[4904]: I0214 11:47:25.310729 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3aaa27d67574612846901271d2b3f13a1f48ee36c1a0d9cc327eefb9ce7de57"} err="failed to get container status \"f3aaa27d67574612846901271d2b3f13a1f48ee36c1a0d9cc327eefb9ce7de57\": rpc error: code = NotFound desc = could not find container \"f3aaa27d67574612846901271d2b3f13a1f48ee36c1a0d9cc327eefb9ce7de57\": container with ID starting with f3aaa27d67574612846901271d2b3f13a1f48ee36c1a0d9cc327eefb9ce7de57 not found: ID does not exist" Feb 14 11:47:25 crc kubenswrapper[4904]: I0214 11:47:25.310746 4904 scope.go:117] "RemoveContainer" containerID="d0d46441c057b598c68f94a0a0009cf756f1760675151e8ca435ef704a57577a" Feb 14 11:47:25 crc kubenswrapper[4904]: E0214 11:47:25.311173 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0d46441c057b598c68f94a0a0009cf756f1760675151e8ca435ef704a57577a\": container with ID starting with d0d46441c057b598c68f94a0a0009cf756f1760675151e8ca435ef704a57577a not found: ID does not exist" containerID="d0d46441c057b598c68f94a0a0009cf756f1760675151e8ca435ef704a57577a" Feb 14 11:47:25 crc kubenswrapper[4904]: I0214 11:47:25.311273 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0d46441c057b598c68f94a0a0009cf756f1760675151e8ca435ef704a57577a"} err="failed to get container status \"d0d46441c057b598c68f94a0a0009cf756f1760675151e8ca435ef704a57577a\": rpc error: code = NotFound desc = could not find container \"d0d46441c057b598c68f94a0a0009cf756f1760675151e8ca435ef704a57577a\": container with ID starting with d0d46441c057b598c68f94a0a0009cf756f1760675151e8ca435ef704a57577a not found: ID does not exist" Feb 14 11:47:25 crc kubenswrapper[4904]: I0214 11:47:25.847810 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13524bff-0982-4b53-9f63-f492a596236e" path="/var/lib/kubelet/pods/13524bff-0982-4b53-9f63-f492a596236e/volumes" Feb 14 11:47:27 crc kubenswrapper[4904]: I0214 11:47:27.837143 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:47:27 crc kubenswrapper[4904]: E0214 11:47:27.837781 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:47:40 crc kubenswrapper[4904]: I0214 11:47:40.566171 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xtlls"] Feb 14 11:47:40 crc kubenswrapper[4904]: E0214 11:47:40.567096 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13524bff-0982-4b53-9f63-f492a596236e" containerName="registry-server" Feb 14 11:47:40 crc kubenswrapper[4904]: I0214 11:47:40.567108 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="13524bff-0982-4b53-9f63-f492a596236e" containerName="registry-server" Feb 14 11:47:40 crc kubenswrapper[4904]: E0214 11:47:40.567118 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13524bff-0982-4b53-9f63-f492a596236e" containerName="extract-content" Feb 14 11:47:40 crc kubenswrapper[4904]: I0214 11:47:40.567124 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="13524bff-0982-4b53-9f63-f492a596236e" containerName="extract-content" Feb 14 11:47:40 crc kubenswrapper[4904]: E0214 11:47:40.567136 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13524bff-0982-4b53-9f63-f492a596236e" containerName="extract-utilities" Feb 14 11:47:40 crc kubenswrapper[4904]: I0214 11:47:40.567142 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="13524bff-0982-4b53-9f63-f492a596236e" containerName="extract-utilities" Feb 14 11:47:40 crc kubenswrapper[4904]: I0214 11:47:40.567314 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="13524bff-0982-4b53-9f63-f492a596236e" containerName="registry-server" Feb 14 11:47:40 crc kubenswrapper[4904]: I0214 11:47:40.568504 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xtlls" Feb 14 11:47:40 crc kubenswrapper[4904]: I0214 11:47:40.583943 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xtlls"] Feb 14 11:47:40 crc kubenswrapper[4904]: I0214 11:47:40.716321 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7-utilities\") pod \"certified-operators-xtlls\" (UID: \"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7\") " pod="openshift-marketplace/certified-operators-xtlls" Feb 14 11:47:40 crc kubenswrapper[4904]: I0214 11:47:40.716394 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7-catalog-content\") pod \"certified-operators-xtlls\" (UID: \"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7\") " pod="openshift-marketplace/certified-operators-xtlls" Feb 14 11:47:40 crc kubenswrapper[4904]: I0214 11:47:40.716543 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz7nx\" (UniqueName: \"kubernetes.io/projected/46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7-kube-api-access-rz7nx\") pod \"certified-operators-xtlls\" (UID: \"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7\") " pod="openshift-marketplace/certified-operators-xtlls" Feb 14 11:47:40 crc kubenswrapper[4904]: I0214 11:47:40.818706 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz7nx\" (UniqueName: \"kubernetes.io/projected/46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7-kube-api-access-rz7nx\") pod \"certified-operators-xtlls\" (UID: \"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7\") " pod="openshift-marketplace/certified-operators-xtlls" Feb 14 11:47:40 crc kubenswrapper[4904]: I0214 11:47:40.818807 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7-utilities\") pod \"certified-operators-xtlls\" (UID: \"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7\") " pod="openshift-marketplace/certified-operators-xtlls" Feb 14 11:47:40 crc kubenswrapper[4904]: I0214 11:47:40.818855 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7-catalog-content\") pod \"certified-operators-xtlls\" (UID: \"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7\") " pod="openshift-marketplace/certified-operators-xtlls" Feb 14 11:47:40 crc kubenswrapper[4904]: I0214 11:47:40.819327 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7-catalog-content\") pod \"certified-operators-xtlls\" (UID: \"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7\") " pod="openshift-marketplace/certified-operators-xtlls" Feb 14 11:47:40 crc kubenswrapper[4904]: I0214 11:47:40.819810 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7-utilities\") pod \"certified-operators-xtlls\" (UID: \"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7\") " pod="openshift-marketplace/certified-operators-xtlls" Feb 14 11:47:40 crc kubenswrapper[4904]: I0214 11:47:40.838721 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz7nx\" (UniqueName: \"kubernetes.io/projected/46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7-kube-api-access-rz7nx\") pod \"certified-operators-xtlls\" (UID: \"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7\") " pod="openshift-marketplace/certified-operators-xtlls" Feb 14 11:47:40 crc kubenswrapper[4904]: I0214 11:47:40.940992 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xtlls" Feb 14 11:47:41 crc kubenswrapper[4904]: I0214 11:47:41.311045 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xtlls"] Feb 14 11:47:41 crc kubenswrapper[4904]: I0214 11:47:41.346074 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xtlls" event={"ID":"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7","Type":"ContainerStarted","Data":"41c8e8166bce8fc029693e57add9fae791346032e7c7e83f6f57965b63794b99"} Feb 14 11:47:41 crc kubenswrapper[4904]: I0214 11:47:41.838248 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:47:41 crc kubenswrapper[4904]: E0214 11:47:41.841361 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:47:42 crc kubenswrapper[4904]: I0214 11:47:42.354811 4904 generic.go:334] "Generic (PLEG): container finished" podID="46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7" containerID="530a827ce06f571042712b8cb0116c7d55ec3546ae0b7aa6963ef0307e94ff29" exitCode=0 Feb 14 11:47:42 crc kubenswrapper[4904]: I0214 11:47:42.354881 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xtlls" event={"ID":"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7","Type":"ContainerDied","Data":"530a827ce06f571042712b8cb0116c7d55ec3546ae0b7aa6963ef0307e94ff29"} Feb 14 11:47:43 crc kubenswrapper[4904]: I0214 11:47:43.363344 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xtlls" event={"ID":"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7","Type":"ContainerStarted","Data":"c5d26932a23272b87e29e3ae5a6c96371f6e56987adb2506f1b629e780f161cd"} Feb 14 11:47:44 crc kubenswrapper[4904]: I0214 11:47:44.380167 4904 generic.go:334] "Generic (PLEG): container finished" podID="46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7" containerID="c5d26932a23272b87e29e3ae5a6c96371f6e56987adb2506f1b629e780f161cd" exitCode=0 Feb 14 11:47:44 crc kubenswrapper[4904]: I0214 11:47:44.380271 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xtlls" event={"ID":"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7","Type":"ContainerDied","Data":"c5d26932a23272b87e29e3ae5a6c96371f6e56987adb2506f1b629e780f161cd"} Feb 14 11:47:45 crc kubenswrapper[4904]: I0214 11:47:45.400480 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xtlls" event={"ID":"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7","Type":"ContainerStarted","Data":"dfd20fcaba158993a6045098f76ea5b3ce605bcb694ea79550eaf2e7f6dd594a"} Feb 14 11:47:45 crc kubenswrapper[4904]: I0214 11:47:45.435715 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xtlls" podStartSLOduration=3.036247358 podStartE2EDuration="5.435698137s" podCreationTimestamp="2026-02-14 11:47:40 +0000 UTC" firstStartedPulling="2026-02-14 11:47:42.356493932 +0000 UTC m=+2253.169258593" lastFinishedPulling="2026-02-14 11:47:44.755944711 +0000 UTC m=+2255.568709372" observedRunningTime="2026-02-14 11:47:45.434629118 +0000 UTC m=+2256.247393769" watchObservedRunningTime="2026-02-14 11:47:45.435698137 +0000 UTC m=+2256.248462798" Feb 14 11:47:50 crc kubenswrapper[4904]: I0214 11:47:50.941966 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xtlls" Feb 14 11:47:50 crc kubenswrapper[4904]: I0214 11:47:50.942562 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xtlls" Feb 14 11:47:50 crc kubenswrapper[4904]: I0214 11:47:50.987756 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xtlls" Feb 14 11:47:51 crc kubenswrapper[4904]: I0214 11:47:51.484850 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xtlls" Feb 14 11:47:51 crc kubenswrapper[4904]: I0214 11:47:51.528009 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xtlls"] Feb 14 11:47:53 crc kubenswrapper[4904]: I0214 11:47:53.457879 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xtlls" podUID="46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7" containerName="registry-server" containerID="cri-o://dfd20fcaba158993a6045098f76ea5b3ce605bcb694ea79550eaf2e7f6dd594a" gracePeriod=2 Feb 14 11:47:53 crc kubenswrapper[4904]: E0214 11:47:53.658992 4904 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46ba3d8d_8c0d_4a09_b84c_c50a9086bbf7.slice/crio-dfd20fcaba158993a6045098f76ea5b3ce605bcb694ea79550eaf2e7f6dd594a.scope\": RecentStats: unable to find data in memory cache]" Feb 14 11:47:53 crc kubenswrapper[4904]: I0214 11:47:53.836373 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:47:53 crc kubenswrapper[4904]: E0214 11:47:53.836712 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.388937 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xtlls" Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.470628 4904 generic.go:334] "Generic (PLEG): container finished" podID="46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7" containerID="dfd20fcaba158993a6045098f76ea5b3ce605bcb694ea79550eaf2e7f6dd594a" exitCode=0 Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.470666 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xtlls" event={"ID":"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7","Type":"ContainerDied","Data":"dfd20fcaba158993a6045098f76ea5b3ce605bcb694ea79550eaf2e7f6dd594a"} Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.470696 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xtlls" event={"ID":"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7","Type":"ContainerDied","Data":"41c8e8166bce8fc029693e57add9fae791346032e7c7e83f6f57965b63794b99"} Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.470695 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xtlls" Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.470713 4904 scope.go:117] "RemoveContainer" containerID="dfd20fcaba158993a6045098f76ea5b3ce605bcb694ea79550eaf2e7f6dd594a" Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.500517 4904 scope.go:117] "RemoveContainer" containerID="c5d26932a23272b87e29e3ae5a6c96371f6e56987adb2506f1b629e780f161cd" Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.515697 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7-utilities\") pod \"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7\" (UID: \"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7\") " Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.515732 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7-catalog-content\") pod \"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7\" (UID: \"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7\") " Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.515927 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rz7nx\" (UniqueName: \"kubernetes.io/projected/46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7-kube-api-access-rz7nx\") pod \"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7\" (UID: \"46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7\") " Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.517013 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7-utilities" (OuterVolumeSpecName: "utilities") pod "46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7" (UID: "46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.524497 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7-kube-api-access-rz7nx" (OuterVolumeSpecName: "kube-api-access-rz7nx") pod "46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7" (UID: "46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7"). InnerVolumeSpecName "kube-api-access-rz7nx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.536944 4904 scope.go:117] "RemoveContainer" containerID="530a827ce06f571042712b8cb0116c7d55ec3546ae0b7aa6963ef0307e94ff29" Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.569023 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7" (UID: "46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.605443 4904 scope.go:117] "RemoveContainer" containerID="dfd20fcaba158993a6045098f76ea5b3ce605bcb694ea79550eaf2e7f6dd594a" Feb 14 11:47:54 crc kubenswrapper[4904]: E0214 11:47:54.606238 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfd20fcaba158993a6045098f76ea5b3ce605bcb694ea79550eaf2e7f6dd594a\": container with ID starting with dfd20fcaba158993a6045098f76ea5b3ce605bcb694ea79550eaf2e7f6dd594a not found: ID does not exist" containerID="dfd20fcaba158993a6045098f76ea5b3ce605bcb694ea79550eaf2e7f6dd594a" Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.606275 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfd20fcaba158993a6045098f76ea5b3ce605bcb694ea79550eaf2e7f6dd594a"} err="failed to get container status \"dfd20fcaba158993a6045098f76ea5b3ce605bcb694ea79550eaf2e7f6dd594a\": rpc error: code = NotFound desc = could not find container \"dfd20fcaba158993a6045098f76ea5b3ce605bcb694ea79550eaf2e7f6dd594a\": container with ID starting with dfd20fcaba158993a6045098f76ea5b3ce605bcb694ea79550eaf2e7f6dd594a not found: ID does not exist" Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.606301 4904 scope.go:117] "RemoveContainer" containerID="c5d26932a23272b87e29e3ae5a6c96371f6e56987adb2506f1b629e780f161cd" Feb 14 11:47:54 crc kubenswrapper[4904]: E0214 11:47:54.606718 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5d26932a23272b87e29e3ae5a6c96371f6e56987adb2506f1b629e780f161cd\": container with ID starting with c5d26932a23272b87e29e3ae5a6c96371f6e56987adb2506f1b629e780f161cd not found: ID does not exist" containerID="c5d26932a23272b87e29e3ae5a6c96371f6e56987adb2506f1b629e780f161cd" Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.606769 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5d26932a23272b87e29e3ae5a6c96371f6e56987adb2506f1b629e780f161cd"} err="failed to get container status \"c5d26932a23272b87e29e3ae5a6c96371f6e56987adb2506f1b629e780f161cd\": rpc error: code = NotFound desc = could not find container \"c5d26932a23272b87e29e3ae5a6c96371f6e56987adb2506f1b629e780f161cd\": container with ID starting with c5d26932a23272b87e29e3ae5a6c96371f6e56987adb2506f1b629e780f161cd not found: ID does not exist" Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.606808 4904 scope.go:117] "RemoveContainer" containerID="530a827ce06f571042712b8cb0116c7d55ec3546ae0b7aa6963ef0307e94ff29" Feb 14 11:47:54 crc kubenswrapper[4904]: E0214 11:47:54.607314 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"530a827ce06f571042712b8cb0116c7d55ec3546ae0b7aa6963ef0307e94ff29\": container with ID starting with 530a827ce06f571042712b8cb0116c7d55ec3546ae0b7aa6963ef0307e94ff29 not found: ID does not exist" containerID="530a827ce06f571042712b8cb0116c7d55ec3546ae0b7aa6963ef0307e94ff29" Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.607343 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"530a827ce06f571042712b8cb0116c7d55ec3546ae0b7aa6963ef0307e94ff29"} err="failed to get container status \"530a827ce06f571042712b8cb0116c7d55ec3546ae0b7aa6963ef0307e94ff29\": rpc error: code = NotFound desc = could not find container \"530a827ce06f571042712b8cb0116c7d55ec3546ae0b7aa6963ef0307e94ff29\": container with ID starting with 530a827ce06f571042712b8cb0116c7d55ec3546ae0b7aa6963ef0307e94ff29 not found: ID does not exist" Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.618388 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rz7nx\" (UniqueName: \"kubernetes.io/projected/46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7-kube-api-access-rz7nx\") on node \"crc\" DevicePath \"\"" Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.618446 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.618456 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.801351 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xtlls"] Feb 14 11:47:54 crc kubenswrapper[4904]: I0214 11:47:54.808896 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xtlls"] Feb 14 11:47:55 crc kubenswrapper[4904]: I0214 11:47:55.846391 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7" path="/var/lib/kubelet/pods/46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7/volumes" Feb 14 11:48:04 crc kubenswrapper[4904]: I0214 11:48:04.836859 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:48:04 crc kubenswrapper[4904]: E0214 11:48:04.837588 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:48:15 crc kubenswrapper[4904]: I0214 11:48:15.836628 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:48:15 crc kubenswrapper[4904]: E0214 11:48:15.837438 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.029642 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr"] Feb 14 11:48:19 crc kubenswrapper[4904]: E0214 11:48:19.030414 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7" containerName="registry-server" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.030434 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7" containerName="registry-server" Feb 14 11:48:19 crc kubenswrapper[4904]: E0214 11:48:19.030472 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7" containerName="extract-content" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.030482 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7" containerName="extract-content" Feb 14 11:48:19 crc kubenswrapper[4904]: E0214 11:48:19.030512 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7" containerName="extract-utilities" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.030521 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7" containerName="extract-utilities" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.030745 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="46ba3d8d-8c0d-4a09-b84c-c50a9086bbf7" containerName="registry-server" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.032463 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.034767 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.036062 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.036266 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.036530 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-48rnz" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.038637 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr"] Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.197966 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c80fc7fa-a761-4701-8ad2-1843b4421958-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr\" (UID: \"c80fc7fa-a761-4701-8ad2-1843b4421958\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.198018 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c80fc7fa-a761-4701-8ad2-1843b4421958-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr\" (UID: \"c80fc7fa-a761-4701-8ad2-1843b4421958\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.198147 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c80fc7fa-a761-4701-8ad2-1843b4421958-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr\" (UID: \"c80fc7fa-a761-4701-8ad2-1843b4421958\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.198174 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smf7n\" (UniqueName: \"kubernetes.io/projected/c80fc7fa-a761-4701-8ad2-1843b4421958-kube-api-access-smf7n\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr\" (UID: \"c80fc7fa-a761-4701-8ad2-1843b4421958\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.299976 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c80fc7fa-a761-4701-8ad2-1843b4421958-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr\" (UID: \"c80fc7fa-a761-4701-8ad2-1843b4421958\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.300298 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c80fc7fa-a761-4701-8ad2-1843b4421958-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr\" (UID: \"c80fc7fa-a761-4701-8ad2-1843b4421958\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.300442 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c80fc7fa-a761-4701-8ad2-1843b4421958-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr\" (UID: \"c80fc7fa-a761-4701-8ad2-1843b4421958\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.300531 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smf7n\" (UniqueName: \"kubernetes.io/projected/c80fc7fa-a761-4701-8ad2-1843b4421958-kube-api-access-smf7n\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr\" (UID: \"c80fc7fa-a761-4701-8ad2-1843b4421958\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.309013 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c80fc7fa-a761-4701-8ad2-1843b4421958-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr\" (UID: \"c80fc7fa-a761-4701-8ad2-1843b4421958\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.309318 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c80fc7fa-a761-4701-8ad2-1843b4421958-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr\" (UID: \"c80fc7fa-a761-4701-8ad2-1843b4421958\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.309686 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c80fc7fa-a761-4701-8ad2-1843b4421958-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr\" (UID: \"c80fc7fa-a761-4701-8ad2-1843b4421958\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.320116 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smf7n\" (UniqueName: \"kubernetes.io/projected/c80fc7fa-a761-4701-8ad2-1843b4421958-kube-api-access-smf7n\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr\" (UID: \"c80fc7fa-a761-4701-8ad2-1843b4421958\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.425406 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.939381 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr"] Feb 14 11:48:19 crc kubenswrapper[4904]: I0214 11:48:19.950876 4904 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 14 11:48:20 crc kubenswrapper[4904]: I0214 11:48:20.702187 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" event={"ID":"c80fc7fa-a761-4701-8ad2-1843b4421958","Type":"ContainerStarted","Data":"7b967bc761c8339c5b0a536d94024117884ffef16f413c100585b7bd92ad8d95"} Feb 14 11:48:20 crc kubenswrapper[4904]: I0214 11:48:20.702449 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" event={"ID":"c80fc7fa-a761-4701-8ad2-1843b4421958","Type":"ContainerStarted","Data":"87c7c362c77e8df93a1aee6d43106c25245b0a5a494502cf23e1b5059b656b7b"} Feb 14 11:48:20 crc kubenswrapper[4904]: I0214 11:48:20.717773 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" podStartSLOduration=1.252281418 podStartE2EDuration="1.717755007s" podCreationTimestamp="2026-02-14 11:48:19 +0000 UTC" firstStartedPulling="2026-02-14 11:48:19.950610844 +0000 UTC m=+2290.763375505" lastFinishedPulling="2026-02-14 11:48:20.416084433 +0000 UTC m=+2291.228849094" observedRunningTime="2026-02-14 11:48:20.716257466 +0000 UTC m=+2291.529022127" watchObservedRunningTime="2026-02-14 11:48:20.717755007 +0000 UTC m=+2291.530519668" Feb 14 11:48:28 crc kubenswrapper[4904]: I0214 11:48:28.837001 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:48:28 crc kubenswrapper[4904]: E0214 11:48:28.837740 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:48:40 crc kubenswrapper[4904]: I0214 11:48:40.838094 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:48:40 crc kubenswrapper[4904]: E0214 11:48:40.839701 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:48:52 crc kubenswrapper[4904]: I0214 11:48:52.836259 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:48:52 crc kubenswrapper[4904]: E0214 11:48:52.836855 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:49:03 crc kubenswrapper[4904]: I0214 11:49:03.836924 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:49:03 crc kubenswrapper[4904]: E0214 11:49:03.837572 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:49:14 crc kubenswrapper[4904]: I0214 11:49:14.836284 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:49:14 crc kubenswrapper[4904]: E0214 11:49:14.837040 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:49:25 crc kubenswrapper[4904]: I0214 11:49:25.836104 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:49:25 crc kubenswrapper[4904]: E0214 11:49:25.837077 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:49:40 crc kubenswrapper[4904]: I0214 11:49:40.836248 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:49:40 crc kubenswrapper[4904]: E0214 11:49:40.836972 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:49:53 crc kubenswrapper[4904]: I0214 11:49:53.836725 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:49:53 crc kubenswrapper[4904]: E0214 11:49:53.838472 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:50:05 crc kubenswrapper[4904]: I0214 11:50:05.836346 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:50:05 crc kubenswrapper[4904]: E0214 11:50:05.837220 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:50:16 crc kubenswrapper[4904]: I0214 11:50:16.836361 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:50:16 crc kubenswrapper[4904]: E0214 11:50:16.837098 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:50:27 crc kubenswrapper[4904]: I0214 11:50:27.836498 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:50:27 crc kubenswrapper[4904]: E0214 11:50:27.837196 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:50:34 crc kubenswrapper[4904]: I0214 11:50:34.795037 4904 generic.go:334] "Generic (PLEG): container finished" podID="c80fc7fa-a761-4701-8ad2-1843b4421958" containerID="7b967bc761c8339c5b0a536d94024117884ffef16f413c100585b7bd92ad8d95" exitCode=2 Feb 14 11:50:34 crc kubenswrapper[4904]: I0214 11:50:34.795524 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" event={"ID":"c80fc7fa-a761-4701-8ad2-1843b4421958","Type":"ContainerDied","Data":"7b967bc761c8339c5b0a536d94024117884ffef16f413c100585b7bd92ad8d95"} Feb 14 11:50:36 crc kubenswrapper[4904]: I0214 11:50:36.189175 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" Feb 14 11:50:36 crc kubenswrapper[4904]: I0214 11:50:36.343087 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smf7n\" (UniqueName: \"kubernetes.io/projected/c80fc7fa-a761-4701-8ad2-1843b4421958-kube-api-access-smf7n\") pod \"c80fc7fa-a761-4701-8ad2-1843b4421958\" (UID: \"c80fc7fa-a761-4701-8ad2-1843b4421958\") " Feb 14 11:50:36 crc kubenswrapper[4904]: I0214 11:50:36.343150 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c80fc7fa-a761-4701-8ad2-1843b4421958-ssh-key-openstack-edpm-ipam\") pod \"c80fc7fa-a761-4701-8ad2-1843b4421958\" (UID: \"c80fc7fa-a761-4701-8ad2-1843b4421958\") " Feb 14 11:50:36 crc kubenswrapper[4904]: I0214 11:50:36.343177 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c80fc7fa-a761-4701-8ad2-1843b4421958-inventory\") pod \"c80fc7fa-a761-4701-8ad2-1843b4421958\" (UID: \"c80fc7fa-a761-4701-8ad2-1843b4421958\") " Feb 14 11:50:36 crc kubenswrapper[4904]: I0214 11:50:36.343405 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c80fc7fa-a761-4701-8ad2-1843b4421958-repo-setup-combined-ca-bundle\") pod \"c80fc7fa-a761-4701-8ad2-1843b4421958\" (UID: \"c80fc7fa-a761-4701-8ad2-1843b4421958\") " Feb 14 11:50:36 crc kubenswrapper[4904]: I0214 11:50:36.353218 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c80fc7fa-a761-4701-8ad2-1843b4421958-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "c80fc7fa-a761-4701-8ad2-1843b4421958" (UID: "c80fc7fa-a761-4701-8ad2-1843b4421958"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:50:36 crc kubenswrapper[4904]: I0214 11:50:36.353253 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c80fc7fa-a761-4701-8ad2-1843b4421958-kube-api-access-smf7n" (OuterVolumeSpecName: "kube-api-access-smf7n") pod "c80fc7fa-a761-4701-8ad2-1843b4421958" (UID: "c80fc7fa-a761-4701-8ad2-1843b4421958"). InnerVolumeSpecName "kube-api-access-smf7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:50:36 crc kubenswrapper[4904]: I0214 11:50:36.370279 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c80fc7fa-a761-4701-8ad2-1843b4421958-inventory" (OuterVolumeSpecName: "inventory") pod "c80fc7fa-a761-4701-8ad2-1843b4421958" (UID: "c80fc7fa-a761-4701-8ad2-1843b4421958"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:50:36 crc kubenswrapper[4904]: I0214 11:50:36.376943 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c80fc7fa-a761-4701-8ad2-1843b4421958-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c80fc7fa-a761-4701-8ad2-1843b4421958" (UID: "c80fc7fa-a761-4701-8ad2-1843b4421958"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:50:36 crc kubenswrapper[4904]: I0214 11:50:36.445632 4904 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c80fc7fa-a761-4701-8ad2-1843b4421958-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:50:36 crc kubenswrapper[4904]: I0214 11:50:36.445670 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smf7n\" (UniqueName: \"kubernetes.io/projected/c80fc7fa-a761-4701-8ad2-1843b4421958-kube-api-access-smf7n\") on node \"crc\" DevicePath \"\"" Feb 14 11:50:36 crc kubenswrapper[4904]: I0214 11:50:36.445680 4904 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c80fc7fa-a761-4701-8ad2-1843b4421958-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 14 11:50:36 crc kubenswrapper[4904]: I0214 11:50:36.445689 4904 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c80fc7fa-a761-4701-8ad2-1843b4421958-inventory\") on node \"crc\" DevicePath \"\"" Feb 14 11:50:36 crc kubenswrapper[4904]: I0214 11:50:36.809391 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" event={"ID":"c80fc7fa-a761-4701-8ad2-1843b4421958","Type":"ContainerDied","Data":"87c7c362c77e8df93a1aee6d43106c25245b0a5a494502cf23e1b5059b656b7b"} Feb 14 11:50:36 crc kubenswrapper[4904]: I0214 11:50:36.809434 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87c7c362c77e8df93a1aee6d43106c25245b0a5a494502cf23e1b5059b656b7b" Feb 14 11:50:36 crc kubenswrapper[4904]: I0214 11:50:36.809701 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr" Feb 14 11:50:42 crc kubenswrapper[4904]: I0214 11:50:42.836918 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:50:42 crc kubenswrapper[4904]: E0214 11:50:42.838521 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:50:55 crc kubenswrapper[4904]: I0214 11:50:55.836307 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:50:55 crc kubenswrapper[4904]: E0214 11:50:55.837125 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:51:10 crc kubenswrapper[4904]: I0214 11:51:10.836127 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:51:10 crc kubenswrapper[4904]: E0214 11:51:10.836989 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:51:25 crc kubenswrapper[4904]: I0214 11:51:25.836421 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:51:25 crc kubenswrapper[4904]: E0214 11:51:25.837198 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:51:30 crc kubenswrapper[4904]: I0214 11:51:30.216709 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-582dc"] Feb 14 11:51:30 crc kubenswrapper[4904]: E0214 11:51:30.217735 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c80fc7fa-a761-4701-8ad2-1843b4421958" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 11:51:30 crc kubenswrapper[4904]: I0214 11:51:30.217752 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="c80fc7fa-a761-4701-8ad2-1843b4421958" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 11:51:30 crc kubenswrapper[4904]: I0214 11:51:30.218035 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="c80fc7fa-a761-4701-8ad2-1843b4421958" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 11:51:30 crc kubenswrapper[4904]: I0214 11:51:30.219669 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-582dc" Feb 14 11:51:30 crc kubenswrapper[4904]: I0214 11:51:30.253020 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-582dc"] Feb 14 11:51:30 crc kubenswrapper[4904]: I0214 11:51:30.338811 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmd5r\" (UniqueName: \"kubernetes.io/projected/03500f1d-50d3-460d-8ce9-590c1b9cb48e-kube-api-access-bmd5r\") pod \"community-operators-582dc\" (UID: \"03500f1d-50d3-460d-8ce9-590c1b9cb48e\") " pod="openshift-marketplace/community-operators-582dc" Feb 14 11:51:30 crc kubenswrapper[4904]: I0214 11:51:30.338916 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03500f1d-50d3-460d-8ce9-590c1b9cb48e-catalog-content\") pod \"community-operators-582dc\" (UID: \"03500f1d-50d3-460d-8ce9-590c1b9cb48e\") " pod="openshift-marketplace/community-operators-582dc" Feb 14 11:51:30 crc kubenswrapper[4904]: I0214 11:51:30.338932 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03500f1d-50d3-460d-8ce9-590c1b9cb48e-utilities\") pod \"community-operators-582dc\" (UID: \"03500f1d-50d3-460d-8ce9-590c1b9cb48e\") " pod="openshift-marketplace/community-operators-582dc" Feb 14 11:51:30 crc kubenswrapper[4904]: I0214 11:51:30.440601 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmd5r\" (UniqueName: \"kubernetes.io/projected/03500f1d-50d3-460d-8ce9-590c1b9cb48e-kube-api-access-bmd5r\") pod \"community-operators-582dc\" (UID: \"03500f1d-50d3-460d-8ce9-590c1b9cb48e\") " pod="openshift-marketplace/community-operators-582dc" Feb 14 11:51:30 crc kubenswrapper[4904]: I0214 11:51:30.440658 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03500f1d-50d3-460d-8ce9-590c1b9cb48e-catalog-content\") pod \"community-operators-582dc\" (UID: \"03500f1d-50d3-460d-8ce9-590c1b9cb48e\") " pod="openshift-marketplace/community-operators-582dc" Feb 14 11:51:30 crc kubenswrapper[4904]: I0214 11:51:30.440678 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03500f1d-50d3-460d-8ce9-590c1b9cb48e-utilities\") pod \"community-operators-582dc\" (UID: \"03500f1d-50d3-460d-8ce9-590c1b9cb48e\") " pod="openshift-marketplace/community-operators-582dc" Feb 14 11:51:30 crc kubenswrapper[4904]: I0214 11:51:30.441163 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03500f1d-50d3-460d-8ce9-590c1b9cb48e-utilities\") pod \"community-operators-582dc\" (UID: \"03500f1d-50d3-460d-8ce9-590c1b9cb48e\") " pod="openshift-marketplace/community-operators-582dc" Feb 14 11:51:30 crc kubenswrapper[4904]: I0214 11:51:30.441234 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03500f1d-50d3-460d-8ce9-590c1b9cb48e-catalog-content\") pod \"community-operators-582dc\" (UID: \"03500f1d-50d3-460d-8ce9-590c1b9cb48e\") " pod="openshift-marketplace/community-operators-582dc" Feb 14 11:51:30 crc kubenswrapper[4904]: I0214 11:51:30.471164 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmd5r\" (UniqueName: \"kubernetes.io/projected/03500f1d-50d3-460d-8ce9-590c1b9cb48e-kube-api-access-bmd5r\") pod \"community-operators-582dc\" (UID: \"03500f1d-50d3-460d-8ce9-590c1b9cb48e\") " pod="openshift-marketplace/community-operators-582dc" Feb 14 11:51:30 crc kubenswrapper[4904]: I0214 11:51:30.548917 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-582dc" Feb 14 11:51:31 crc kubenswrapper[4904]: I0214 11:51:31.105857 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-582dc"] Feb 14 11:51:31 crc kubenswrapper[4904]: I0214 11:51:31.248182 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-582dc" event={"ID":"03500f1d-50d3-460d-8ce9-590c1b9cb48e","Type":"ContainerStarted","Data":"5b3c951cdbf75227a3c2b8d4092919f064cad14ca520e34eb088c69de2ed17aa"} Feb 14 11:51:32 crc kubenswrapper[4904]: I0214 11:51:32.258883 4904 generic.go:334] "Generic (PLEG): container finished" podID="03500f1d-50d3-460d-8ce9-590c1b9cb48e" containerID="1a0c8631bbcace0ef7477592110d7100bb99380c44ac78dd1bedd3bc1fd41b00" exitCode=0 Feb 14 11:51:32 crc kubenswrapper[4904]: I0214 11:51:32.258954 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-582dc" event={"ID":"03500f1d-50d3-460d-8ce9-590c1b9cb48e","Type":"ContainerDied","Data":"1a0c8631bbcace0ef7477592110d7100bb99380c44ac78dd1bedd3bc1fd41b00"} Feb 14 11:51:33 crc kubenswrapper[4904]: I0214 11:51:33.269289 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-582dc" event={"ID":"03500f1d-50d3-460d-8ce9-590c1b9cb48e","Type":"ContainerStarted","Data":"924c84408cfae18623d6a48814484d42d900d6f3ffded34ef868cca868a222d7"} Feb 14 11:51:34 crc kubenswrapper[4904]: I0214 11:51:34.279763 4904 generic.go:334] "Generic (PLEG): container finished" podID="03500f1d-50d3-460d-8ce9-590c1b9cb48e" containerID="924c84408cfae18623d6a48814484d42d900d6f3ffded34ef868cca868a222d7" exitCode=0 Feb 14 11:51:34 crc kubenswrapper[4904]: I0214 11:51:34.279848 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-582dc" event={"ID":"03500f1d-50d3-460d-8ce9-590c1b9cb48e","Type":"ContainerDied","Data":"924c84408cfae18623d6a48814484d42d900d6f3ffded34ef868cca868a222d7"} Feb 14 11:51:35 crc kubenswrapper[4904]: I0214 11:51:35.293417 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-582dc" event={"ID":"03500f1d-50d3-460d-8ce9-590c1b9cb48e","Type":"ContainerStarted","Data":"5f546049f932b983267075ea4125203e63f76d5af7660629892d384ccc77b5ce"} Feb 14 11:51:35 crc kubenswrapper[4904]: I0214 11:51:35.320666 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-582dc" podStartSLOduration=2.904232959 podStartE2EDuration="5.3206314s" podCreationTimestamp="2026-02-14 11:51:30 +0000 UTC" firstStartedPulling="2026-02-14 11:51:32.260713185 +0000 UTC m=+2483.073477846" lastFinishedPulling="2026-02-14 11:51:34.677111626 +0000 UTC m=+2485.489876287" observedRunningTime="2026-02-14 11:51:35.310564264 +0000 UTC m=+2486.123328965" watchObservedRunningTime="2026-02-14 11:51:35.3206314 +0000 UTC m=+2486.133396091" Feb 14 11:51:40 crc kubenswrapper[4904]: I0214 11:51:40.549164 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-582dc" Feb 14 11:51:40 crc kubenswrapper[4904]: I0214 11:51:40.549777 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-582dc" Feb 14 11:51:40 crc kubenswrapper[4904]: I0214 11:51:40.623556 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-582dc" Feb 14 11:51:40 crc kubenswrapper[4904]: I0214 11:51:40.675583 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fv6pm"] Feb 14 11:51:40 crc kubenswrapper[4904]: I0214 11:51:40.680369 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fv6pm" Feb 14 11:51:40 crc kubenswrapper[4904]: I0214 11:51:40.718100 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fv6pm"] Feb 14 11:51:40 crc kubenswrapper[4904]: I0214 11:51:40.760368 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de04a444-acc7-4ca4-b227-186fdbcd4b11-utilities\") pod \"redhat-marketplace-fv6pm\" (UID: \"de04a444-acc7-4ca4-b227-186fdbcd4b11\") " pod="openshift-marketplace/redhat-marketplace-fv6pm" Feb 14 11:51:40 crc kubenswrapper[4904]: I0214 11:51:40.760554 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de04a444-acc7-4ca4-b227-186fdbcd4b11-catalog-content\") pod \"redhat-marketplace-fv6pm\" (UID: \"de04a444-acc7-4ca4-b227-186fdbcd4b11\") " pod="openshift-marketplace/redhat-marketplace-fv6pm" Feb 14 11:51:40 crc kubenswrapper[4904]: I0214 11:51:40.760642 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcd6z\" (UniqueName: \"kubernetes.io/projected/de04a444-acc7-4ca4-b227-186fdbcd4b11-kube-api-access-zcd6z\") pod \"redhat-marketplace-fv6pm\" (UID: \"de04a444-acc7-4ca4-b227-186fdbcd4b11\") " pod="openshift-marketplace/redhat-marketplace-fv6pm" Feb 14 11:51:40 crc kubenswrapper[4904]: I0214 11:51:40.836780 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:51:40 crc kubenswrapper[4904]: E0214 11:51:40.837193 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:51:40 crc kubenswrapper[4904]: I0214 11:51:40.862208 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de04a444-acc7-4ca4-b227-186fdbcd4b11-catalog-content\") pod \"redhat-marketplace-fv6pm\" (UID: \"de04a444-acc7-4ca4-b227-186fdbcd4b11\") " pod="openshift-marketplace/redhat-marketplace-fv6pm" Feb 14 11:51:40 crc kubenswrapper[4904]: I0214 11:51:40.862280 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcd6z\" (UniqueName: \"kubernetes.io/projected/de04a444-acc7-4ca4-b227-186fdbcd4b11-kube-api-access-zcd6z\") pod \"redhat-marketplace-fv6pm\" (UID: \"de04a444-acc7-4ca4-b227-186fdbcd4b11\") " pod="openshift-marketplace/redhat-marketplace-fv6pm" Feb 14 11:51:40 crc kubenswrapper[4904]: I0214 11:51:40.862333 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de04a444-acc7-4ca4-b227-186fdbcd4b11-utilities\") pod \"redhat-marketplace-fv6pm\" (UID: \"de04a444-acc7-4ca4-b227-186fdbcd4b11\") " pod="openshift-marketplace/redhat-marketplace-fv6pm" Feb 14 11:51:40 crc kubenswrapper[4904]: I0214 11:51:40.862675 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de04a444-acc7-4ca4-b227-186fdbcd4b11-catalog-content\") pod \"redhat-marketplace-fv6pm\" (UID: \"de04a444-acc7-4ca4-b227-186fdbcd4b11\") " pod="openshift-marketplace/redhat-marketplace-fv6pm" Feb 14 11:51:40 crc kubenswrapper[4904]: I0214 11:51:40.863062 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de04a444-acc7-4ca4-b227-186fdbcd4b11-utilities\") pod \"redhat-marketplace-fv6pm\" (UID: \"de04a444-acc7-4ca4-b227-186fdbcd4b11\") " pod="openshift-marketplace/redhat-marketplace-fv6pm" Feb 14 11:51:40 crc kubenswrapper[4904]: I0214 11:51:40.882992 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcd6z\" (UniqueName: \"kubernetes.io/projected/de04a444-acc7-4ca4-b227-186fdbcd4b11-kube-api-access-zcd6z\") pod \"redhat-marketplace-fv6pm\" (UID: \"de04a444-acc7-4ca4-b227-186fdbcd4b11\") " pod="openshift-marketplace/redhat-marketplace-fv6pm" Feb 14 11:51:41 crc kubenswrapper[4904]: I0214 11:51:41.004111 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fv6pm" Feb 14 11:51:41 crc kubenswrapper[4904]: I0214 11:51:41.391878 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-582dc" Feb 14 11:51:41 crc kubenswrapper[4904]: I0214 11:51:41.479851 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fv6pm"] Feb 14 11:51:41 crc kubenswrapper[4904]: W0214 11:51:41.486591 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde04a444_acc7_4ca4_b227_186fdbcd4b11.slice/crio-86e43ffc599ba1c7383b7d56261327b2a2e90608f11f2eb00407f15876ffa2cf WatchSource:0}: Error finding container 86e43ffc599ba1c7383b7d56261327b2a2e90608f11f2eb00407f15876ffa2cf: Status 404 returned error can't find the container with id 86e43ffc599ba1c7383b7d56261327b2a2e90608f11f2eb00407f15876ffa2cf Feb 14 11:51:42 crc kubenswrapper[4904]: I0214 11:51:42.345316 4904 generic.go:334] "Generic (PLEG): container finished" podID="de04a444-acc7-4ca4-b227-186fdbcd4b11" containerID="58965d9de8cd9c2cb70fd29c5e7613a79874eabec7469f49e32842baa2390b7b" exitCode=0 Feb 14 11:51:42 crc kubenswrapper[4904]: I0214 11:51:42.346175 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fv6pm" event={"ID":"de04a444-acc7-4ca4-b227-186fdbcd4b11","Type":"ContainerDied","Data":"58965d9de8cd9c2cb70fd29c5e7613a79874eabec7469f49e32842baa2390b7b"} Feb 14 11:51:42 crc kubenswrapper[4904]: I0214 11:51:42.346214 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fv6pm" event={"ID":"de04a444-acc7-4ca4-b227-186fdbcd4b11","Type":"ContainerStarted","Data":"86e43ffc599ba1c7383b7d56261327b2a2e90608f11f2eb00407f15876ffa2cf"} Feb 14 11:51:42 crc kubenswrapper[4904]: I0214 11:51:42.821432 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-582dc"] Feb 14 11:51:43 crc kubenswrapper[4904]: I0214 11:51:43.356745 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fv6pm" event={"ID":"de04a444-acc7-4ca4-b227-186fdbcd4b11","Type":"ContainerStarted","Data":"53acacb3fbe391ba93f2cab2e5850fb5053bd5f767ac076b2b084184f0e309c8"} Feb 14 11:51:43 crc kubenswrapper[4904]: I0214 11:51:43.356892 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-582dc" podUID="03500f1d-50d3-460d-8ce9-590c1b9cb48e" containerName="registry-server" containerID="cri-o://5f546049f932b983267075ea4125203e63f76d5af7660629892d384ccc77b5ce" gracePeriod=2 Feb 14 11:51:43 crc kubenswrapper[4904]: I0214 11:51:43.803504 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-582dc" Feb 14 11:51:43 crc kubenswrapper[4904]: I0214 11:51:43.821530 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmd5r\" (UniqueName: \"kubernetes.io/projected/03500f1d-50d3-460d-8ce9-590c1b9cb48e-kube-api-access-bmd5r\") pod \"03500f1d-50d3-460d-8ce9-590c1b9cb48e\" (UID: \"03500f1d-50d3-460d-8ce9-590c1b9cb48e\") " Feb 14 11:51:43 crc kubenswrapper[4904]: I0214 11:51:43.821678 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03500f1d-50d3-460d-8ce9-590c1b9cb48e-utilities\") pod \"03500f1d-50d3-460d-8ce9-590c1b9cb48e\" (UID: \"03500f1d-50d3-460d-8ce9-590c1b9cb48e\") " Feb 14 11:51:43 crc kubenswrapper[4904]: I0214 11:51:43.821768 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03500f1d-50d3-460d-8ce9-590c1b9cb48e-catalog-content\") pod \"03500f1d-50d3-460d-8ce9-590c1b9cb48e\" (UID: \"03500f1d-50d3-460d-8ce9-590c1b9cb48e\") " Feb 14 11:51:43 crc kubenswrapper[4904]: I0214 11:51:43.824392 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03500f1d-50d3-460d-8ce9-590c1b9cb48e-utilities" (OuterVolumeSpecName: "utilities") pod "03500f1d-50d3-460d-8ce9-590c1b9cb48e" (UID: "03500f1d-50d3-460d-8ce9-590c1b9cb48e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:51:43 crc kubenswrapper[4904]: I0214 11:51:43.834316 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03500f1d-50d3-460d-8ce9-590c1b9cb48e-kube-api-access-bmd5r" (OuterVolumeSpecName: "kube-api-access-bmd5r") pod "03500f1d-50d3-460d-8ce9-590c1b9cb48e" (UID: "03500f1d-50d3-460d-8ce9-590c1b9cb48e"). InnerVolumeSpecName "kube-api-access-bmd5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:51:43 crc kubenswrapper[4904]: I0214 11:51:43.923983 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmd5r\" (UniqueName: \"kubernetes.io/projected/03500f1d-50d3-460d-8ce9-590c1b9cb48e-kube-api-access-bmd5r\") on node \"crc\" DevicePath \"\"" Feb 14 11:51:43 crc kubenswrapper[4904]: I0214 11:51:43.924373 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03500f1d-50d3-460d-8ce9-590c1b9cb48e-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:51:44 crc kubenswrapper[4904]: I0214 11:51:44.370987 4904 generic.go:334] "Generic (PLEG): container finished" podID="03500f1d-50d3-460d-8ce9-590c1b9cb48e" containerID="5f546049f932b983267075ea4125203e63f76d5af7660629892d384ccc77b5ce" exitCode=0 Feb 14 11:51:44 crc kubenswrapper[4904]: I0214 11:51:44.371049 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-582dc" Feb 14 11:51:44 crc kubenswrapper[4904]: I0214 11:51:44.371048 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-582dc" event={"ID":"03500f1d-50d3-460d-8ce9-590c1b9cb48e","Type":"ContainerDied","Data":"5f546049f932b983267075ea4125203e63f76d5af7660629892d384ccc77b5ce"} Feb 14 11:51:44 crc kubenswrapper[4904]: I0214 11:51:44.372235 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-582dc" event={"ID":"03500f1d-50d3-460d-8ce9-590c1b9cb48e","Type":"ContainerDied","Data":"5b3c951cdbf75227a3c2b8d4092919f064cad14ca520e34eb088c69de2ed17aa"} Feb 14 11:51:44 crc kubenswrapper[4904]: I0214 11:51:44.372276 4904 scope.go:117] "RemoveContainer" containerID="5f546049f932b983267075ea4125203e63f76d5af7660629892d384ccc77b5ce" Feb 14 11:51:44 crc kubenswrapper[4904]: I0214 11:51:44.389937 4904 scope.go:117] "RemoveContainer" containerID="924c84408cfae18623d6a48814484d42d900d6f3ffded34ef868cca868a222d7" Feb 14 11:51:44 crc kubenswrapper[4904]: I0214 11:51:44.412489 4904 scope.go:117] "RemoveContainer" containerID="1a0c8631bbcace0ef7477592110d7100bb99380c44ac78dd1bedd3bc1fd41b00" Feb 14 11:51:44 crc kubenswrapper[4904]: I0214 11:51:44.468402 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03500f1d-50d3-460d-8ce9-590c1b9cb48e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03500f1d-50d3-460d-8ce9-590c1b9cb48e" (UID: "03500f1d-50d3-460d-8ce9-590c1b9cb48e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:51:44 crc kubenswrapper[4904]: I0214 11:51:44.508224 4904 scope.go:117] "RemoveContainer" containerID="5f546049f932b983267075ea4125203e63f76d5af7660629892d384ccc77b5ce" Feb 14 11:51:44 crc kubenswrapper[4904]: E0214 11:51:44.508807 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f546049f932b983267075ea4125203e63f76d5af7660629892d384ccc77b5ce\": container with ID starting with 5f546049f932b983267075ea4125203e63f76d5af7660629892d384ccc77b5ce not found: ID does not exist" containerID="5f546049f932b983267075ea4125203e63f76d5af7660629892d384ccc77b5ce" Feb 14 11:51:44 crc kubenswrapper[4904]: I0214 11:51:44.508895 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f546049f932b983267075ea4125203e63f76d5af7660629892d384ccc77b5ce"} err="failed to get container status \"5f546049f932b983267075ea4125203e63f76d5af7660629892d384ccc77b5ce\": rpc error: code = NotFound desc = could not find container \"5f546049f932b983267075ea4125203e63f76d5af7660629892d384ccc77b5ce\": container with ID starting with 5f546049f932b983267075ea4125203e63f76d5af7660629892d384ccc77b5ce not found: ID does not exist" Feb 14 11:51:44 crc kubenswrapper[4904]: I0214 11:51:44.508922 4904 scope.go:117] "RemoveContainer" containerID="924c84408cfae18623d6a48814484d42d900d6f3ffded34ef868cca868a222d7" Feb 14 11:51:44 crc kubenswrapper[4904]: E0214 11:51:44.509274 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"924c84408cfae18623d6a48814484d42d900d6f3ffded34ef868cca868a222d7\": container with ID starting with 924c84408cfae18623d6a48814484d42d900d6f3ffded34ef868cca868a222d7 not found: ID does not exist" containerID="924c84408cfae18623d6a48814484d42d900d6f3ffded34ef868cca868a222d7" Feb 14 11:51:44 crc kubenswrapper[4904]: I0214 11:51:44.509376 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"924c84408cfae18623d6a48814484d42d900d6f3ffded34ef868cca868a222d7"} err="failed to get container status \"924c84408cfae18623d6a48814484d42d900d6f3ffded34ef868cca868a222d7\": rpc error: code = NotFound desc = could not find container \"924c84408cfae18623d6a48814484d42d900d6f3ffded34ef868cca868a222d7\": container with ID starting with 924c84408cfae18623d6a48814484d42d900d6f3ffded34ef868cca868a222d7 not found: ID does not exist" Feb 14 11:51:44 crc kubenswrapper[4904]: I0214 11:51:44.509470 4904 scope.go:117] "RemoveContainer" containerID="1a0c8631bbcace0ef7477592110d7100bb99380c44ac78dd1bedd3bc1fd41b00" Feb 14 11:51:44 crc kubenswrapper[4904]: E0214 11:51:44.510237 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a0c8631bbcace0ef7477592110d7100bb99380c44ac78dd1bedd3bc1fd41b00\": container with ID starting with 1a0c8631bbcace0ef7477592110d7100bb99380c44ac78dd1bedd3bc1fd41b00 not found: ID does not exist" containerID="1a0c8631bbcace0ef7477592110d7100bb99380c44ac78dd1bedd3bc1fd41b00" Feb 14 11:51:44 crc kubenswrapper[4904]: I0214 11:51:44.510272 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a0c8631bbcace0ef7477592110d7100bb99380c44ac78dd1bedd3bc1fd41b00"} err="failed to get container status \"1a0c8631bbcace0ef7477592110d7100bb99380c44ac78dd1bedd3bc1fd41b00\": rpc error: code = NotFound desc = could not find container \"1a0c8631bbcace0ef7477592110d7100bb99380c44ac78dd1bedd3bc1fd41b00\": container with ID starting with 1a0c8631bbcace0ef7477592110d7100bb99380c44ac78dd1bedd3bc1fd41b00 not found: ID does not exist" Feb 14 11:51:44 crc kubenswrapper[4904]: I0214 11:51:44.539024 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03500f1d-50d3-460d-8ce9-590c1b9cb48e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:51:44 crc kubenswrapper[4904]: I0214 11:51:44.706363 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-582dc"] Feb 14 11:51:44 crc kubenswrapper[4904]: I0214 11:51:44.714056 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-582dc"] Feb 14 11:51:45 crc kubenswrapper[4904]: I0214 11:51:45.392166 4904 generic.go:334] "Generic (PLEG): container finished" podID="de04a444-acc7-4ca4-b227-186fdbcd4b11" containerID="53acacb3fbe391ba93f2cab2e5850fb5053bd5f767ac076b2b084184f0e309c8" exitCode=0 Feb 14 11:51:45 crc kubenswrapper[4904]: I0214 11:51:45.392256 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fv6pm" event={"ID":"de04a444-acc7-4ca4-b227-186fdbcd4b11","Type":"ContainerDied","Data":"53acacb3fbe391ba93f2cab2e5850fb5053bd5f767ac076b2b084184f0e309c8"} Feb 14 11:51:45 crc kubenswrapper[4904]: I0214 11:51:45.845983 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03500f1d-50d3-460d-8ce9-590c1b9cb48e" path="/var/lib/kubelet/pods/03500f1d-50d3-460d-8ce9-590c1b9cb48e/volumes" Feb 14 11:51:46 crc kubenswrapper[4904]: I0214 11:51:46.416742 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fv6pm" event={"ID":"de04a444-acc7-4ca4-b227-186fdbcd4b11","Type":"ContainerStarted","Data":"05f238ee4aad73ad0d7825f7ac825666a32fe9aafc6c8366477998fba9e38304"} Feb 14 11:51:46 crc kubenswrapper[4904]: I0214 11:51:46.447691 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fv6pm" podStartSLOduration=2.897752655 podStartE2EDuration="6.447671294s" podCreationTimestamp="2026-02-14 11:51:40 +0000 UTC" firstStartedPulling="2026-02-14 11:51:42.348138815 +0000 UTC m=+2493.160903476" lastFinishedPulling="2026-02-14 11:51:45.898057454 +0000 UTC m=+2496.710822115" observedRunningTime="2026-02-14 11:51:46.438259825 +0000 UTC m=+2497.251024486" watchObservedRunningTime="2026-02-14 11:51:46.447671294 +0000 UTC m=+2497.260435965" Feb 14 11:51:51 crc kubenswrapper[4904]: I0214 11:51:51.005112 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fv6pm" Feb 14 11:51:51 crc kubenswrapper[4904]: I0214 11:51:51.005748 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fv6pm" Feb 14 11:51:51 crc kubenswrapper[4904]: I0214 11:51:51.053886 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fv6pm" Feb 14 11:51:51 crc kubenswrapper[4904]: I0214 11:51:51.532506 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fv6pm" Feb 14 11:51:51 crc kubenswrapper[4904]: I0214 11:51:51.600155 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fv6pm"] Feb 14 11:51:53 crc kubenswrapper[4904]: I0214 11:51:53.469498 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fv6pm" podUID="de04a444-acc7-4ca4-b227-186fdbcd4b11" containerName="registry-server" containerID="cri-o://05f238ee4aad73ad0d7825f7ac825666a32fe9aafc6c8366477998fba9e38304" gracePeriod=2 Feb 14 11:51:53 crc kubenswrapper[4904]: I0214 11:51:53.861670 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fv6pm" Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.044898 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de04a444-acc7-4ca4-b227-186fdbcd4b11-catalog-content\") pod \"de04a444-acc7-4ca4-b227-186fdbcd4b11\" (UID: \"de04a444-acc7-4ca4-b227-186fdbcd4b11\") " Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.045057 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcd6z\" (UniqueName: \"kubernetes.io/projected/de04a444-acc7-4ca4-b227-186fdbcd4b11-kube-api-access-zcd6z\") pod \"de04a444-acc7-4ca4-b227-186fdbcd4b11\" (UID: \"de04a444-acc7-4ca4-b227-186fdbcd4b11\") " Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.045182 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de04a444-acc7-4ca4-b227-186fdbcd4b11-utilities\") pod \"de04a444-acc7-4ca4-b227-186fdbcd4b11\" (UID: \"de04a444-acc7-4ca4-b227-186fdbcd4b11\") " Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.046022 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de04a444-acc7-4ca4-b227-186fdbcd4b11-utilities" (OuterVolumeSpecName: "utilities") pod "de04a444-acc7-4ca4-b227-186fdbcd4b11" (UID: "de04a444-acc7-4ca4-b227-186fdbcd4b11"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.046450 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de04a444-acc7-4ca4-b227-186fdbcd4b11-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.051138 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de04a444-acc7-4ca4-b227-186fdbcd4b11-kube-api-access-zcd6z" (OuterVolumeSpecName: "kube-api-access-zcd6z") pod "de04a444-acc7-4ca4-b227-186fdbcd4b11" (UID: "de04a444-acc7-4ca4-b227-186fdbcd4b11"). InnerVolumeSpecName "kube-api-access-zcd6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.072040 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de04a444-acc7-4ca4-b227-186fdbcd4b11-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de04a444-acc7-4ca4-b227-186fdbcd4b11" (UID: "de04a444-acc7-4ca4-b227-186fdbcd4b11"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.147429 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcd6z\" (UniqueName: \"kubernetes.io/projected/de04a444-acc7-4ca4-b227-186fdbcd4b11-kube-api-access-zcd6z\") on node \"crc\" DevicePath \"\"" Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.147459 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de04a444-acc7-4ca4-b227-186fdbcd4b11-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.478690 4904 generic.go:334] "Generic (PLEG): container finished" podID="de04a444-acc7-4ca4-b227-186fdbcd4b11" containerID="05f238ee4aad73ad0d7825f7ac825666a32fe9aafc6c8366477998fba9e38304" exitCode=0 Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.478775 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fv6pm" Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.478797 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fv6pm" event={"ID":"de04a444-acc7-4ca4-b227-186fdbcd4b11","Type":"ContainerDied","Data":"05f238ee4aad73ad0d7825f7ac825666a32fe9aafc6c8366477998fba9e38304"} Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.479796 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fv6pm" event={"ID":"de04a444-acc7-4ca4-b227-186fdbcd4b11","Type":"ContainerDied","Data":"86e43ffc599ba1c7383b7d56261327b2a2e90608f11f2eb00407f15876ffa2cf"} Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.479826 4904 scope.go:117] "RemoveContainer" containerID="05f238ee4aad73ad0d7825f7ac825666a32fe9aafc6c8366477998fba9e38304" Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.509431 4904 scope.go:117] "RemoveContainer" containerID="53acacb3fbe391ba93f2cab2e5850fb5053bd5f767ac076b2b084184f0e309c8" Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.524302 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fv6pm"] Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.533171 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fv6pm"] Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.546057 4904 scope.go:117] "RemoveContainer" containerID="58965d9de8cd9c2cb70fd29c5e7613a79874eabec7469f49e32842baa2390b7b" Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.575533 4904 scope.go:117] "RemoveContainer" containerID="05f238ee4aad73ad0d7825f7ac825666a32fe9aafc6c8366477998fba9e38304" Feb 14 11:51:54 crc kubenswrapper[4904]: E0214 11:51:54.576214 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05f238ee4aad73ad0d7825f7ac825666a32fe9aafc6c8366477998fba9e38304\": container with ID starting with 05f238ee4aad73ad0d7825f7ac825666a32fe9aafc6c8366477998fba9e38304 not found: ID does not exist" containerID="05f238ee4aad73ad0d7825f7ac825666a32fe9aafc6c8366477998fba9e38304" Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.576267 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05f238ee4aad73ad0d7825f7ac825666a32fe9aafc6c8366477998fba9e38304"} err="failed to get container status \"05f238ee4aad73ad0d7825f7ac825666a32fe9aafc6c8366477998fba9e38304\": rpc error: code = NotFound desc = could not find container \"05f238ee4aad73ad0d7825f7ac825666a32fe9aafc6c8366477998fba9e38304\": container with ID starting with 05f238ee4aad73ad0d7825f7ac825666a32fe9aafc6c8366477998fba9e38304 not found: ID does not exist" Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.576307 4904 scope.go:117] "RemoveContainer" containerID="53acacb3fbe391ba93f2cab2e5850fb5053bd5f767ac076b2b084184f0e309c8" Feb 14 11:51:54 crc kubenswrapper[4904]: E0214 11:51:54.576757 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53acacb3fbe391ba93f2cab2e5850fb5053bd5f767ac076b2b084184f0e309c8\": container with ID starting with 53acacb3fbe391ba93f2cab2e5850fb5053bd5f767ac076b2b084184f0e309c8 not found: ID does not exist" containerID="53acacb3fbe391ba93f2cab2e5850fb5053bd5f767ac076b2b084184f0e309c8" Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.576799 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53acacb3fbe391ba93f2cab2e5850fb5053bd5f767ac076b2b084184f0e309c8"} err="failed to get container status \"53acacb3fbe391ba93f2cab2e5850fb5053bd5f767ac076b2b084184f0e309c8\": rpc error: code = NotFound desc = could not find container \"53acacb3fbe391ba93f2cab2e5850fb5053bd5f767ac076b2b084184f0e309c8\": container with ID starting with 53acacb3fbe391ba93f2cab2e5850fb5053bd5f767ac076b2b084184f0e309c8 not found: ID does not exist" Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.576850 4904 scope.go:117] "RemoveContainer" containerID="58965d9de8cd9c2cb70fd29c5e7613a79874eabec7469f49e32842baa2390b7b" Feb 14 11:51:54 crc kubenswrapper[4904]: E0214 11:51:54.578542 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58965d9de8cd9c2cb70fd29c5e7613a79874eabec7469f49e32842baa2390b7b\": container with ID starting with 58965d9de8cd9c2cb70fd29c5e7613a79874eabec7469f49e32842baa2390b7b not found: ID does not exist" containerID="58965d9de8cd9c2cb70fd29c5e7613a79874eabec7469f49e32842baa2390b7b" Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.578564 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58965d9de8cd9c2cb70fd29c5e7613a79874eabec7469f49e32842baa2390b7b"} err="failed to get container status \"58965d9de8cd9c2cb70fd29c5e7613a79874eabec7469f49e32842baa2390b7b\": rpc error: code = NotFound desc = could not find container \"58965d9de8cd9c2cb70fd29c5e7613a79874eabec7469f49e32842baa2390b7b\": container with ID starting with 58965d9de8cd9c2cb70fd29c5e7613a79874eabec7469f49e32842baa2390b7b not found: ID does not exist" Feb 14 11:51:54 crc kubenswrapper[4904]: I0214 11:51:54.836504 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:51:54 crc kubenswrapper[4904]: E0214 11:51:54.836716 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:51:55 crc kubenswrapper[4904]: I0214 11:51:55.846524 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de04a444-acc7-4ca4-b227-186fdbcd4b11" path="/var/lib/kubelet/pods/de04a444-acc7-4ca4-b227-186fdbcd4b11/volumes" Feb 14 11:52:05 crc kubenswrapper[4904]: I0214 11:52:05.837118 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:52:05 crc kubenswrapper[4904]: E0214 11:52:05.837945 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:52:16 crc kubenswrapper[4904]: I0214 11:52:16.836661 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:52:17 crc kubenswrapper[4904]: I0214 11:52:17.691275 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"a4d54d5738bdee810566472d9ce65fa4b7fb204a38e6994280faa872ccdf04e5"} Feb 14 11:54:46 crc kubenswrapper[4904]: I0214 11:54:46.383082 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:54:46 crc kubenswrapper[4904]: I0214 11:54:46.383557 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:55:16 crc kubenswrapper[4904]: I0214 11:55:16.383004 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:55:16 crc kubenswrapper[4904]: I0214 11:55:16.383565 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:55:46 crc kubenswrapper[4904]: I0214 11:55:46.383373 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:55:46 crc kubenswrapper[4904]: I0214 11:55:46.383895 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:55:46 crc kubenswrapper[4904]: I0214 11:55:46.383958 4904 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:55:46 crc kubenswrapper[4904]: I0214 11:55:46.384950 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a4d54d5738bdee810566472d9ce65fa4b7fb204a38e6994280faa872ccdf04e5"} pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 14 11:55:46 crc kubenswrapper[4904]: I0214 11:55:46.385106 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" containerID="cri-o://a4d54d5738bdee810566472d9ce65fa4b7fb204a38e6994280faa872ccdf04e5" gracePeriod=600 Feb 14 11:55:46 crc kubenswrapper[4904]: I0214 11:55:46.798004 4904 generic.go:334] "Generic (PLEG): container finished" podID="7790d276-2506-4641-8c22-67e6441417f0" containerID="a4d54d5738bdee810566472d9ce65fa4b7fb204a38e6994280faa872ccdf04e5" exitCode=0 Feb 14 11:55:46 crc kubenswrapper[4904]: I0214 11:55:46.798080 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerDied","Data":"a4d54d5738bdee810566472d9ce65fa4b7fb204a38e6994280faa872ccdf04e5"} Feb 14 11:55:46 crc kubenswrapper[4904]: I0214 11:55:46.798348 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067"} Feb 14 11:55:46 crc kubenswrapper[4904]: I0214 11:55:46.798376 4904 scope.go:117] "RemoveContainer" containerID="2303070eb86d23e663d21f4e18aedc73b5f01d0b060d091347d81fc156cbe4fb" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.030238 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7"] Feb 14 11:55:54 crc kubenswrapper[4904]: E0214 11:55:54.031983 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de04a444-acc7-4ca4-b227-186fdbcd4b11" containerName="extract-content" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.032093 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="de04a444-acc7-4ca4-b227-186fdbcd4b11" containerName="extract-content" Feb 14 11:55:54 crc kubenswrapper[4904]: E0214 11:55:54.032165 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03500f1d-50d3-460d-8ce9-590c1b9cb48e" containerName="registry-server" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.032222 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="03500f1d-50d3-460d-8ce9-590c1b9cb48e" containerName="registry-server" Feb 14 11:55:54 crc kubenswrapper[4904]: E0214 11:55:54.032289 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de04a444-acc7-4ca4-b227-186fdbcd4b11" containerName="registry-server" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.032367 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="de04a444-acc7-4ca4-b227-186fdbcd4b11" containerName="registry-server" Feb 14 11:55:54 crc kubenswrapper[4904]: E0214 11:55:54.032431 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03500f1d-50d3-460d-8ce9-590c1b9cb48e" containerName="extract-content" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.032490 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="03500f1d-50d3-460d-8ce9-590c1b9cb48e" containerName="extract-content" Feb 14 11:55:54 crc kubenswrapper[4904]: E0214 11:55:54.032552 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03500f1d-50d3-460d-8ce9-590c1b9cb48e" containerName="extract-utilities" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.032611 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="03500f1d-50d3-460d-8ce9-590c1b9cb48e" containerName="extract-utilities" Feb 14 11:55:54 crc kubenswrapper[4904]: E0214 11:55:54.032670 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de04a444-acc7-4ca4-b227-186fdbcd4b11" containerName="extract-utilities" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.032727 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="de04a444-acc7-4ca4-b227-186fdbcd4b11" containerName="extract-utilities" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.032989 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="03500f1d-50d3-460d-8ce9-590c1b9cb48e" containerName="registry-server" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.033073 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="de04a444-acc7-4ca4-b227-186fdbcd4b11" containerName="registry-server" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.033768 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.036050 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.036164 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.037611 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-48rnz" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.038328 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.040624 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7"] Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.086263 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7\" (UID: \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.086315 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngp69\" (UniqueName: \"kubernetes.io/projected/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-kube-api-access-ngp69\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7\" (UID: \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.086382 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7\" (UID: \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.086528 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7\" (UID: \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.188485 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7\" (UID: \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.188535 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngp69\" (UniqueName: \"kubernetes.io/projected/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-kube-api-access-ngp69\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7\" (UID: \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.188620 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7\" (UID: \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.188649 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7\" (UID: \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.194445 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7\" (UID: \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.194491 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7\" (UID: \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.198956 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7\" (UID: \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.207705 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngp69\" (UniqueName: \"kubernetes.io/projected/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-kube-api-access-ngp69\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7\" (UID: \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.364315 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.865294 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7"] Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.874085 4904 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 14 11:55:54 crc kubenswrapper[4904]: I0214 11:55:54.886264 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" event={"ID":"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb","Type":"ContainerStarted","Data":"91f0a86f442be18d11397ff5920ed4d912a4da7049312cf7be1d294f8f3f0ff2"} Feb 14 11:55:55 crc kubenswrapper[4904]: I0214 11:55:55.898965 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" event={"ID":"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb","Type":"ContainerStarted","Data":"48cc9fec7258bb9957fc0f5ec03b2abdcd33c8e76126aa87d68c86741f8a6b58"} Feb 14 11:55:55 crc kubenswrapper[4904]: I0214 11:55:55.916889 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" podStartSLOduration=1.461401076 podStartE2EDuration="1.916869734s" podCreationTimestamp="2026-02-14 11:55:54 +0000 UTC" firstStartedPulling="2026-02-14 11:55:54.873843887 +0000 UTC m=+2745.686608548" lastFinishedPulling="2026-02-14 11:55:55.329312545 +0000 UTC m=+2746.142077206" observedRunningTime="2026-02-14 11:55:55.915301802 +0000 UTC m=+2746.728066473" watchObservedRunningTime="2026-02-14 11:55:55.916869734 +0000 UTC m=+2746.729634395" Feb 14 11:57:34 crc kubenswrapper[4904]: I0214 11:57:34.839112 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l92xz"] Feb 14 11:57:34 crc kubenswrapper[4904]: I0214 11:57:34.843359 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l92xz" Feb 14 11:57:34 crc kubenswrapper[4904]: I0214 11:57:34.860062 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l92xz"] Feb 14 11:57:34 crc kubenswrapper[4904]: I0214 11:57:34.985736 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8cb20e1-81f2-4c7c-b0a7-24b18c938e25-catalog-content\") pod \"redhat-operators-l92xz\" (UID: \"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25\") " pod="openshift-marketplace/redhat-operators-l92xz" Feb 14 11:57:34 crc kubenswrapper[4904]: I0214 11:57:34.986114 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8cb20e1-81f2-4c7c-b0a7-24b18c938e25-utilities\") pod \"redhat-operators-l92xz\" (UID: \"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25\") " pod="openshift-marketplace/redhat-operators-l92xz" Feb 14 11:57:34 crc kubenswrapper[4904]: I0214 11:57:34.986222 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5vgr\" (UniqueName: \"kubernetes.io/projected/a8cb20e1-81f2-4c7c-b0a7-24b18c938e25-kube-api-access-v5vgr\") pod \"redhat-operators-l92xz\" (UID: \"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25\") " pod="openshift-marketplace/redhat-operators-l92xz" Feb 14 11:57:35 crc kubenswrapper[4904]: I0214 11:57:35.087542 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5vgr\" (UniqueName: \"kubernetes.io/projected/a8cb20e1-81f2-4c7c-b0a7-24b18c938e25-kube-api-access-v5vgr\") pod \"redhat-operators-l92xz\" (UID: \"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25\") " pod="openshift-marketplace/redhat-operators-l92xz" Feb 14 11:57:35 crc kubenswrapper[4904]: I0214 11:57:35.087862 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8cb20e1-81f2-4c7c-b0a7-24b18c938e25-catalog-content\") pod \"redhat-operators-l92xz\" (UID: \"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25\") " pod="openshift-marketplace/redhat-operators-l92xz" Feb 14 11:57:35 crc kubenswrapper[4904]: I0214 11:57:35.087918 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8cb20e1-81f2-4c7c-b0a7-24b18c938e25-utilities\") pod \"redhat-operators-l92xz\" (UID: \"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25\") " pod="openshift-marketplace/redhat-operators-l92xz" Feb 14 11:57:35 crc kubenswrapper[4904]: I0214 11:57:35.088370 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8cb20e1-81f2-4c7c-b0a7-24b18c938e25-catalog-content\") pod \"redhat-operators-l92xz\" (UID: \"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25\") " pod="openshift-marketplace/redhat-operators-l92xz" Feb 14 11:57:35 crc kubenswrapper[4904]: I0214 11:57:35.088390 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8cb20e1-81f2-4c7c-b0a7-24b18c938e25-utilities\") pod \"redhat-operators-l92xz\" (UID: \"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25\") " pod="openshift-marketplace/redhat-operators-l92xz" Feb 14 11:57:35 crc kubenswrapper[4904]: I0214 11:57:35.106060 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5vgr\" (UniqueName: \"kubernetes.io/projected/a8cb20e1-81f2-4c7c-b0a7-24b18c938e25-kube-api-access-v5vgr\") pod \"redhat-operators-l92xz\" (UID: \"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25\") " pod="openshift-marketplace/redhat-operators-l92xz" Feb 14 11:57:35 crc kubenswrapper[4904]: I0214 11:57:35.168449 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l92xz" Feb 14 11:57:35 crc kubenswrapper[4904]: I0214 11:57:35.692243 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l92xz"] Feb 14 11:57:36 crc kubenswrapper[4904]: I0214 11:57:36.699562 4904 generic.go:334] "Generic (PLEG): container finished" podID="a8cb20e1-81f2-4c7c-b0a7-24b18c938e25" containerID="3378de83b43e794d6ee8a77ad61156e612e3b020e3589355bc75702df73fb91b" exitCode=0 Feb 14 11:57:36 crc kubenswrapper[4904]: I0214 11:57:36.699618 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l92xz" event={"ID":"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25","Type":"ContainerDied","Data":"3378de83b43e794d6ee8a77ad61156e612e3b020e3589355bc75702df73fb91b"} Feb 14 11:57:36 crc kubenswrapper[4904]: I0214 11:57:36.699976 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l92xz" event={"ID":"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25","Type":"ContainerStarted","Data":"bf2e222d0419c457c6ed6fead8adc54efd265feacda134c6f63196c442aea41e"} Feb 14 11:57:37 crc kubenswrapper[4904]: I0214 11:57:37.711915 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l92xz" event={"ID":"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25","Type":"ContainerStarted","Data":"bb49e4c2f54723ef40a48a647131c0b3cd84b40f6fc5bbe4a9e94e2afc7393b9"} Feb 14 11:57:42 crc kubenswrapper[4904]: I0214 11:57:42.757774 4904 generic.go:334] "Generic (PLEG): container finished" podID="a8cb20e1-81f2-4c7c-b0a7-24b18c938e25" containerID="bb49e4c2f54723ef40a48a647131c0b3cd84b40f6fc5bbe4a9e94e2afc7393b9" exitCode=0 Feb 14 11:57:42 crc kubenswrapper[4904]: I0214 11:57:42.761563 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l92xz" event={"ID":"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25","Type":"ContainerDied","Data":"bb49e4c2f54723ef40a48a647131c0b3cd84b40f6fc5bbe4a9e94e2afc7393b9"} Feb 14 11:57:43 crc kubenswrapper[4904]: I0214 11:57:43.775968 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l92xz" event={"ID":"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25","Type":"ContainerStarted","Data":"69ec67a0495972fda4e7c2ad2ed7eb75d98ac663d802f122f50c1c32cbf8e6e9"} Feb 14 11:57:43 crc kubenswrapper[4904]: I0214 11:57:43.821373 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l92xz" podStartSLOduration=3.386062518 podStartE2EDuration="9.821344896s" podCreationTimestamp="2026-02-14 11:57:34 +0000 UTC" firstStartedPulling="2026-02-14 11:57:36.70152223 +0000 UTC m=+2847.514286891" lastFinishedPulling="2026-02-14 11:57:43.136804598 +0000 UTC m=+2853.949569269" observedRunningTime="2026-02-14 11:57:43.810596002 +0000 UTC m=+2854.623360683" watchObservedRunningTime="2026-02-14 11:57:43.821344896 +0000 UTC m=+2854.634109577" Feb 14 11:57:45 crc kubenswrapper[4904]: I0214 11:57:45.168818 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l92xz" Feb 14 11:57:45 crc kubenswrapper[4904]: I0214 11:57:45.169178 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l92xz" Feb 14 11:57:46 crc kubenswrapper[4904]: I0214 11:57:46.215435 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l92xz" podUID="a8cb20e1-81f2-4c7c-b0a7-24b18c938e25" containerName="registry-server" probeResult="failure" output=< Feb 14 11:57:46 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 11:57:46 crc kubenswrapper[4904]: > Feb 14 11:57:46 crc kubenswrapper[4904]: I0214 11:57:46.382862 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:57:46 crc kubenswrapper[4904]: I0214 11:57:46.382962 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:57:56 crc kubenswrapper[4904]: I0214 11:57:56.212721 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l92xz" podUID="a8cb20e1-81f2-4c7c-b0a7-24b18c938e25" containerName="registry-server" probeResult="failure" output=< Feb 14 11:57:56 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 11:57:56 crc kubenswrapper[4904]: > Feb 14 11:58:05 crc kubenswrapper[4904]: I0214 11:58:05.238957 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l92xz" Feb 14 11:58:05 crc kubenswrapper[4904]: I0214 11:58:05.309223 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l92xz" Feb 14 11:58:06 crc kubenswrapper[4904]: I0214 11:58:06.039079 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l92xz"] Feb 14 11:58:06 crc kubenswrapper[4904]: I0214 11:58:06.973480 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l92xz" podUID="a8cb20e1-81f2-4c7c-b0a7-24b18c938e25" containerName="registry-server" containerID="cri-o://69ec67a0495972fda4e7c2ad2ed7eb75d98ac663d802f122f50c1c32cbf8e6e9" gracePeriod=2 Feb 14 11:58:07 crc kubenswrapper[4904]: I0214 11:58:07.414063 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l92xz" Feb 14 11:58:07 crc kubenswrapper[4904]: I0214 11:58:07.562163 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5vgr\" (UniqueName: \"kubernetes.io/projected/a8cb20e1-81f2-4c7c-b0a7-24b18c938e25-kube-api-access-v5vgr\") pod \"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25\" (UID: \"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25\") " Feb 14 11:58:07 crc kubenswrapper[4904]: I0214 11:58:07.562231 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8cb20e1-81f2-4c7c-b0a7-24b18c938e25-utilities\") pod \"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25\" (UID: \"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25\") " Feb 14 11:58:07 crc kubenswrapper[4904]: I0214 11:58:07.562399 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8cb20e1-81f2-4c7c-b0a7-24b18c938e25-catalog-content\") pod \"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25\" (UID: \"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25\") " Feb 14 11:58:07 crc kubenswrapper[4904]: I0214 11:58:07.563384 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8cb20e1-81f2-4c7c-b0a7-24b18c938e25-utilities" (OuterVolumeSpecName: "utilities") pod "a8cb20e1-81f2-4c7c-b0a7-24b18c938e25" (UID: "a8cb20e1-81f2-4c7c-b0a7-24b18c938e25"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:58:07 crc kubenswrapper[4904]: I0214 11:58:07.572346 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8cb20e1-81f2-4c7c-b0a7-24b18c938e25-kube-api-access-v5vgr" (OuterVolumeSpecName: "kube-api-access-v5vgr") pod "a8cb20e1-81f2-4c7c-b0a7-24b18c938e25" (UID: "a8cb20e1-81f2-4c7c-b0a7-24b18c938e25"). InnerVolumeSpecName "kube-api-access-v5vgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:58:07 crc kubenswrapper[4904]: I0214 11:58:07.664626 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5vgr\" (UniqueName: \"kubernetes.io/projected/a8cb20e1-81f2-4c7c-b0a7-24b18c938e25-kube-api-access-v5vgr\") on node \"crc\" DevicePath \"\"" Feb 14 11:58:07 crc kubenswrapper[4904]: I0214 11:58:07.664668 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8cb20e1-81f2-4c7c-b0a7-24b18c938e25-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 11:58:07 crc kubenswrapper[4904]: I0214 11:58:07.693101 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8cb20e1-81f2-4c7c-b0a7-24b18c938e25-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a8cb20e1-81f2-4c7c-b0a7-24b18c938e25" (UID: "a8cb20e1-81f2-4c7c-b0a7-24b18c938e25"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 11:58:07 crc kubenswrapper[4904]: I0214 11:58:07.766570 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8cb20e1-81f2-4c7c-b0a7-24b18c938e25-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 11:58:07 crc kubenswrapper[4904]: I0214 11:58:07.981785 4904 generic.go:334] "Generic (PLEG): container finished" podID="a8cb20e1-81f2-4c7c-b0a7-24b18c938e25" containerID="69ec67a0495972fda4e7c2ad2ed7eb75d98ac663d802f122f50c1c32cbf8e6e9" exitCode=0 Feb 14 11:58:07 crc kubenswrapper[4904]: I0214 11:58:07.981823 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l92xz" event={"ID":"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25","Type":"ContainerDied","Data":"69ec67a0495972fda4e7c2ad2ed7eb75d98ac663d802f122f50c1c32cbf8e6e9"} Feb 14 11:58:07 crc kubenswrapper[4904]: I0214 11:58:07.981861 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l92xz" event={"ID":"a8cb20e1-81f2-4c7c-b0a7-24b18c938e25","Type":"ContainerDied","Data":"bf2e222d0419c457c6ed6fead8adc54efd265feacda134c6f63196c442aea41e"} Feb 14 11:58:07 crc kubenswrapper[4904]: I0214 11:58:07.981878 4904 scope.go:117] "RemoveContainer" containerID="69ec67a0495972fda4e7c2ad2ed7eb75d98ac663d802f122f50c1c32cbf8e6e9" Feb 14 11:58:07 crc kubenswrapper[4904]: I0214 11:58:07.981832 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l92xz" Feb 14 11:58:08 crc kubenswrapper[4904]: I0214 11:58:08.012621 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l92xz"] Feb 14 11:58:08 crc kubenswrapper[4904]: I0214 11:58:08.013795 4904 scope.go:117] "RemoveContainer" containerID="bb49e4c2f54723ef40a48a647131c0b3cd84b40f6fc5bbe4a9e94e2afc7393b9" Feb 14 11:58:08 crc kubenswrapper[4904]: I0214 11:58:08.021029 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l92xz"] Feb 14 11:58:08 crc kubenswrapper[4904]: E0214 11:58:08.038552 4904 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8cb20e1_81f2_4c7c_b0a7_24b18c938e25.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8cb20e1_81f2_4c7c_b0a7_24b18c938e25.slice/crio-bf2e222d0419c457c6ed6fead8adc54efd265feacda134c6f63196c442aea41e\": RecentStats: unable to find data in memory cache]" Feb 14 11:58:08 crc kubenswrapper[4904]: I0214 11:58:08.043567 4904 scope.go:117] "RemoveContainer" containerID="3378de83b43e794d6ee8a77ad61156e612e3b020e3589355bc75702df73fb91b" Feb 14 11:58:08 crc kubenswrapper[4904]: I0214 11:58:08.080745 4904 scope.go:117] "RemoveContainer" containerID="69ec67a0495972fda4e7c2ad2ed7eb75d98ac663d802f122f50c1c32cbf8e6e9" Feb 14 11:58:08 crc kubenswrapper[4904]: E0214 11:58:08.081333 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69ec67a0495972fda4e7c2ad2ed7eb75d98ac663d802f122f50c1c32cbf8e6e9\": container with ID starting with 69ec67a0495972fda4e7c2ad2ed7eb75d98ac663d802f122f50c1c32cbf8e6e9 not found: ID does not exist" containerID="69ec67a0495972fda4e7c2ad2ed7eb75d98ac663d802f122f50c1c32cbf8e6e9" Feb 14 11:58:08 crc kubenswrapper[4904]: I0214 11:58:08.081371 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69ec67a0495972fda4e7c2ad2ed7eb75d98ac663d802f122f50c1c32cbf8e6e9"} err="failed to get container status \"69ec67a0495972fda4e7c2ad2ed7eb75d98ac663d802f122f50c1c32cbf8e6e9\": rpc error: code = NotFound desc = could not find container \"69ec67a0495972fda4e7c2ad2ed7eb75d98ac663d802f122f50c1c32cbf8e6e9\": container with ID starting with 69ec67a0495972fda4e7c2ad2ed7eb75d98ac663d802f122f50c1c32cbf8e6e9 not found: ID does not exist" Feb 14 11:58:08 crc kubenswrapper[4904]: I0214 11:58:08.081398 4904 scope.go:117] "RemoveContainer" containerID="bb49e4c2f54723ef40a48a647131c0b3cd84b40f6fc5bbe4a9e94e2afc7393b9" Feb 14 11:58:08 crc kubenswrapper[4904]: E0214 11:58:08.081632 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb49e4c2f54723ef40a48a647131c0b3cd84b40f6fc5bbe4a9e94e2afc7393b9\": container with ID starting with bb49e4c2f54723ef40a48a647131c0b3cd84b40f6fc5bbe4a9e94e2afc7393b9 not found: ID does not exist" containerID="bb49e4c2f54723ef40a48a647131c0b3cd84b40f6fc5bbe4a9e94e2afc7393b9" Feb 14 11:58:08 crc kubenswrapper[4904]: I0214 11:58:08.081663 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb49e4c2f54723ef40a48a647131c0b3cd84b40f6fc5bbe4a9e94e2afc7393b9"} err="failed to get container status \"bb49e4c2f54723ef40a48a647131c0b3cd84b40f6fc5bbe4a9e94e2afc7393b9\": rpc error: code = NotFound desc = could not find container \"bb49e4c2f54723ef40a48a647131c0b3cd84b40f6fc5bbe4a9e94e2afc7393b9\": container with ID starting with bb49e4c2f54723ef40a48a647131c0b3cd84b40f6fc5bbe4a9e94e2afc7393b9 not found: ID does not exist" Feb 14 11:58:08 crc kubenswrapper[4904]: I0214 11:58:08.081678 4904 scope.go:117] "RemoveContainer" containerID="3378de83b43e794d6ee8a77ad61156e612e3b020e3589355bc75702df73fb91b" Feb 14 11:58:08 crc kubenswrapper[4904]: E0214 11:58:08.081887 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3378de83b43e794d6ee8a77ad61156e612e3b020e3589355bc75702df73fb91b\": container with ID starting with 3378de83b43e794d6ee8a77ad61156e612e3b020e3589355bc75702df73fb91b not found: ID does not exist" containerID="3378de83b43e794d6ee8a77ad61156e612e3b020e3589355bc75702df73fb91b" Feb 14 11:58:08 crc kubenswrapper[4904]: I0214 11:58:08.081907 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3378de83b43e794d6ee8a77ad61156e612e3b020e3589355bc75702df73fb91b"} err="failed to get container status \"3378de83b43e794d6ee8a77ad61156e612e3b020e3589355bc75702df73fb91b\": rpc error: code = NotFound desc = could not find container \"3378de83b43e794d6ee8a77ad61156e612e3b020e3589355bc75702df73fb91b\": container with ID starting with 3378de83b43e794d6ee8a77ad61156e612e3b020e3589355bc75702df73fb91b not found: ID does not exist" Feb 14 11:58:08 crc kubenswrapper[4904]: I0214 11:58:08.990123 4904 generic.go:334] "Generic (PLEG): container finished" podID="6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb" containerID="48cc9fec7258bb9957fc0f5ec03b2abdcd33c8e76126aa87d68c86741f8a6b58" exitCode=2 Feb 14 11:58:08 crc kubenswrapper[4904]: I0214 11:58:08.990181 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" event={"ID":"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb","Type":"ContainerDied","Data":"48cc9fec7258bb9957fc0f5ec03b2abdcd33c8e76126aa87d68c86741f8a6b58"} Feb 14 11:58:09 crc kubenswrapper[4904]: I0214 11:58:09.862026 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8cb20e1-81f2-4c7c-b0a7-24b18c938e25" path="/var/lib/kubelet/pods/a8cb20e1-81f2-4c7c-b0a7-24b18c938e25/volumes" Feb 14 11:58:10 crc kubenswrapper[4904]: I0214 11:58:10.392883 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" Feb 14 11:58:10 crc kubenswrapper[4904]: I0214 11:58:10.516369 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-repo-setup-combined-ca-bundle\") pod \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\" (UID: \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\") " Feb 14 11:58:10 crc kubenswrapper[4904]: I0214 11:58:10.516650 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-ssh-key-openstack-edpm-ipam\") pod \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\" (UID: \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\") " Feb 14 11:58:10 crc kubenswrapper[4904]: I0214 11:58:10.516741 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngp69\" (UniqueName: \"kubernetes.io/projected/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-kube-api-access-ngp69\") pod \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\" (UID: \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\") " Feb 14 11:58:10 crc kubenswrapper[4904]: I0214 11:58:10.516810 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-inventory\") pod \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\" (UID: \"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb\") " Feb 14 11:58:10 crc kubenswrapper[4904]: I0214 11:58:10.526059 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb" (UID: "6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:58:10 crc kubenswrapper[4904]: I0214 11:58:10.526089 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-kube-api-access-ngp69" (OuterVolumeSpecName: "kube-api-access-ngp69") pod "6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb" (UID: "6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb"). InnerVolumeSpecName "kube-api-access-ngp69". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 11:58:10 crc kubenswrapper[4904]: I0214 11:58:10.540406 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-inventory" (OuterVolumeSpecName: "inventory") pod "6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb" (UID: "6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:58:10 crc kubenswrapper[4904]: I0214 11:58:10.542374 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb" (UID: "6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 11:58:10 crc kubenswrapper[4904]: I0214 11:58:10.618193 4904 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 11:58:10 crc kubenswrapper[4904]: I0214 11:58:10.618226 4904 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 14 11:58:10 crc kubenswrapper[4904]: I0214 11:58:10.618238 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngp69\" (UniqueName: \"kubernetes.io/projected/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-kube-api-access-ngp69\") on node \"crc\" DevicePath \"\"" Feb 14 11:58:10 crc kubenswrapper[4904]: I0214 11:58:10.618248 4904 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb-inventory\") on node \"crc\" DevicePath \"\"" Feb 14 11:58:11 crc kubenswrapper[4904]: I0214 11:58:11.009407 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" event={"ID":"6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb","Type":"ContainerDied","Data":"91f0a86f442be18d11397ff5920ed4d912a4da7049312cf7be1d294f8f3f0ff2"} Feb 14 11:58:11 crc kubenswrapper[4904]: I0214 11:58:11.009686 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91f0a86f442be18d11397ff5920ed4d912a4da7049312cf7be1d294f8f3f0ff2" Feb 14 11:58:11 crc kubenswrapper[4904]: I0214 11:58:11.009444 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7" Feb 14 11:58:16 crc kubenswrapper[4904]: I0214 11:58:16.383536 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:58:16 crc kubenswrapper[4904]: I0214 11:58:16.384225 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:58:46 crc kubenswrapper[4904]: I0214 11:58:46.382675 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 11:58:46 crc kubenswrapper[4904]: I0214 11:58:46.383362 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 11:58:46 crc kubenswrapper[4904]: I0214 11:58:46.383426 4904 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 11:58:46 crc kubenswrapper[4904]: I0214 11:58:46.384591 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067"} pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 14 11:58:46 crc kubenswrapper[4904]: I0214 11:58:46.384776 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" containerID="cri-o://0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" gracePeriod=600 Feb 14 11:58:46 crc kubenswrapper[4904]: E0214 11:58:46.508374 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:58:47 crc kubenswrapper[4904]: I0214 11:58:47.372684 4904 generic.go:334] "Generic (PLEG): container finished" podID="7790d276-2506-4641-8c22-67e6441417f0" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" exitCode=0 Feb 14 11:58:47 crc kubenswrapper[4904]: I0214 11:58:47.372742 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerDied","Data":"0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067"} Feb 14 11:58:47 crc kubenswrapper[4904]: I0214 11:58:47.372795 4904 scope.go:117] "RemoveContainer" containerID="a4d54d5738bdee810566472d9ce65fa4b7fb204a38e6994280faa872ccdf04e5" Feb 14 11:58:47 crc kubenswrapper[4904]: I0214 11:58:47.373620 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 11:58:47 crc kubenswrapper[4904]: E0214 11:58:47.374018 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:59:01 crc kubenswrapper[4904]: I0214 11:59:01.837772 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 11:59:01 crc kubenswrapper[4904]: E0214 11:59:01.839404 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:59:16 crc kubenswrapper[4904]: I0214 11:59:16.836482 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 11:59:16 crc kubenswrapper[4904]: E0214 11:59:16.837275 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:59:29 crc kubenswrapper[4904]: I0214 11:59:29.843273 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 11:59:29 crc kubenswrapper[4904]: E0214 11:59:29.844955 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:59:41 crc kubenswrapper[4904]: I0214 11:59:41.836897 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 11:59:41 crc kubenswrapper[4904]: E0214 11:59:41.837669 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 11:59:53 crc kubenswrapper[4904]: I0214 11:59:53.836396 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 11:59:53 crc kubenswrapper[4904]: E0214 11:59:53.837979 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.145983 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517840-ntbnl"] Feb 14 12:00:00 crc kubenswrapper[4904]: E0214 12:00:00.147239 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8cb20e1-81f2-4c7c-b0a7-24b18c938e25" containerName="extract-content" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.147258 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8cb20e1-81f2-4c7c-b0a7-24b18c938e25" containerName="extract-content" Feb 14 12:00:00 crc kubenswrapper[4904]: E0214 12:00:00.147271 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.147281 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 12:00:00 crc kubenswrapper[4904]: E0214 12:00:00.147297 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8cb20e1-81f2-4c7c-b0a7-24b18c938e25" containerName="registry-server" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.147304 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8cb20e1-81f2-4c7c-b0a7-24b18c938e25" containerName="registry-server" Feb 14 12:00:00 crc kubenswrapper[4904]: E0214 12:00:00.147323 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8cb20e1-81f2-4c7c-b0a7-24b18c938e25" containerName="extract-utilities" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.147331 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8cb20e1-81f2-4c7c-b0a7-24b18c938e25" containerName="extract-utilities" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.147562 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.147589 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8cb20e1-81f2-4c7c-b0a7-24b18c938e25" containerName="registry-server" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.148375 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517840-ntbnl" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.153021 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.153357 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.236616 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517840-ntbnl"] Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.297368 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16ca79af-7c49-4b29-96d5-87dca2fb8ead-secret-volume\") pod \"collect-profiles-29517840-ntbnl\" (UID: \"16ca79af-7c49-4b29-96d5-87dca2fb8ead\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517840-ntbnl" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.297428 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16ca79af-7c49-4b29-96d5-87dca2fb8ead-config-volume\") pod \"collect-profiles-29517840-ntbnl\" (UID: \"16ca79af-7c49-4b29-96d5-87dca2fb8ead\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517840-ntbnl" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.297490 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vkkg\" (UniqueName: \"kubernetes.io/projected/16ca79af-7c49-4b29-96d5-87dca2fb8ead-kube-api-access-9vkkg\") pod \"collect-profiles-29517840-ntbnl\" (UID: \"16ca79af-7c49-4b29-96d5-87dca2fb8ead\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517840-ntbnl" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.399747 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16ca79af-7c49-4b29-96d5-87dca2fb8ead-secret-volume\") pod \"collect-profiles-29517840-ntbnl\" (UID: \"16ca79af-7c49-4b29-96d5-87dca2fb8ead\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517840-ntbnl" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.399805 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16ca79af-7c49-4b29-96d5-87dca2fb8ead-config-volume\") pod \"collect-profiles-29517840-ntbnl\" (UID: \"16ca79af-7c49-4b29-96d5-87dca2fb8ead\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517840-ntbnl" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.399910 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vkkg\" (UniqueName: \"kubernetes.io/projected/16ca79af-7c49-4b29-96d5-87dca2fb8ead-kube-api-access-9vkkg\") pod \"collect-profiles-29517840-ntbnl\" (UID: \"16ca79af-7c49-4b29-96d5-87dca2fb8ead\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517840-ntbnl" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.402636 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16ca79af-7c49-4b29-96d5-87dca2fb8ead-config-volume\") pod \"collect-profiles-29517840-ntbnl\" (UID: \"16ca79af-7c49-4b29-96d5-87dca2fb8ead\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517840-ntbnl" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.406592 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16ca79af-7c49-4b29-96d5-87dca2fb8ead-secret-volume\") pod \"collect-profiles-29517840-ntbnl\" (UID: \"16ca79af-7c49-4b29-96d5-87dca2fb8ead\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517840-ntbnl" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.419958 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vkkg\" (UniqueName: \"kubernetes.io/projected/16ca79af-7c49-4b29-96d5-87dca2fb8ead-kube-api-access-9vkkg\") pod \"collect-profiles-29517840-ntbnl\" (UID: \"16ca79af-7c49-4b29-96d5-87dca2fb8ead\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517840-ntbnl" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.466305 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517840-ntbnl" Feb 14 12:00:00 crc kubenswrapper[4904]: I0214 12:00:00.939870 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517840-ntbnl"] Feb 14 12:00:01 crc kubenswrapper[4904]: I0214 12:00:01.009265 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517840-ntbnl" event={"ID":"16ca79af-7c49-4b29-96d5-87dca2fb8ead","Type":"ContainerStarted","Data":"eca213fbc1f7acd233b5d105be752a05ad159b0e8d847286b9f2919b113d91b8"} Feb 14 12:00:02 crc kubenswrapper[4904]: I0214 12:00:02.018058 4904 generic.go:334] "Generic (PLEG): container finished" podID="16ca79af-7c49-4b29-96d5-87dca2fb8ead" containerID="15f2a8a14b6705d069bdfa6875ee5aed3ebb3f234cf828a110015b66d15224ca" exitCode=0 Feb 14 12:00:02 crc kubenswrapper[4904]: I0214 12:00:02.018117 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517840-ntbnl" event={"ID":"16ca79af-7c49-4b29-96d5-87dca2fb8ead","Type":"ContainerDied","Data":"15f2a8a14b6705d069bdfa6875ee5aed3ebb3f234cf828a110015b66d15224ca"} Feb 14 12:00:03 crc kubenswrapper[4904]: I0214 12:00:03.354213 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517840-ntbnl" Feb 14 12:00:03 crc kubenswrapper[4904]: I0214 12:00:03.476524 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16ca79af-7c49-4b29-96d5-87dca2fb8ead-config-volume\") pod \"16ca79af-7c49-4b29-96d5-87dca2fb8ead\" (UID: \"16ca79af-7c49-4b29-96d5-87dca2fb8ead\") " Feb 14 12:00:03 crc kubenswrapper[4904]: I0214 12:00:03.476672 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vkkg\" (UniqueName: \"kubernetes.io/projected/16ca79af-7c49-4b29-96d5-87dca2fb8ead-kube-api-access-9vkkg\") pod \"16ca79af-7c49-4b29-96d5-87dca2fb8ead\" (UID: \"16ca79af-7c49-4b29-96d5-87dca2fb8ead\") " Feb 14 12:00:03 crc kubenswrapper[4904]: I0214 12:00:03.476872 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16ca79af-7c49-4b29-96d5-87dca2fb8ead-secret-volume\") pod \"16ca79af-7c49-4b29-96d5-87dca2fb8ead\" (UID: \"16ca79af-7c49-4b29-96d5-87dca2fb8ead\") " Feb 14 12:00:03 crc kubenswrapper[4904]: I0214 12:00:03.477451 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16ca79af-7c49-4b29-96d5-87dca2fb8ead-config-volume" (OuterVolumeSpecName: "config-volume") pod "16ca79af-7c49-4b29-96d5-87dca2fb8ead" (UID: "16ca79af-7c49-4b29-96d5-87dca2fb8ead"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 12:00:03 crc kubenswrapper[4904]: I0214 12:00:03.481907 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16ca79af-7c49-4b29-96d5-87dca2fb8ead-kube-api-access-9vkkg" (OuterVolumeSpecName: "kube-api-access-9vkkg") pod "16ca79af-7c49-4b29-96d5-87dca2fb8ead" (UID: "16ca79af-7c49-4b29-96d5-87dca2fb8ead"). InnerVolumeSpecName "kube-api-access-9vkkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:00:03 crc kubenswrapper[4904]: I0214 12:00:03.481951 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16ca79af-7c49-4b29-96d5-87dca2fb8ead-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "16ca79af-7c49-4b29-96d5-87dca2fb8ead" (UID: "16ca79af-7c49-4b29-96d5-87dca2fb8ead"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 12:00:03 crc kubenswrapper[4904]: I0214 12:00:03.578673 4904 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16ca79af-7c49-4b29-96d5-87dca2fb8ead-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 14 12:00:03 crc kubenswrapper[4904]: I0214 12:00:03.579074 4904 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16ca79af-7c49-4b29-96d5-87dca2fb8ead-config-volume\") on node \"crc\" DevicePath \"\"" Feb 14 12:00:03 crc kubenswrapper[4904]: I0214 12:00:03.579087 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vkkg\" (UniqueName: \"kubernetes.io/projected/16ca79af-7c49-4b29-96d5-87dca2fb8ead-kube-api-access-9vkkg\") on node \"crc\" DevicePath \"\"" Feb 14 12:00:04 crc kubenswrapper[4904]: I0214 12:00:04.034858 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517840-ntbnl" event={"ID":"16ca79af-7c49-4b29-96d5-87dca2fb8ead","Type":"ContainerDied","Data":"eca213fbc1f7acd233b5d105be752a05ad159b0e8d847286b9f2919b113d91b8"} Feb 14 12:00:04 crc kubenswrapper[4904]: I0214 12:00:04.034900 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eca213fbc1f7acd233b5d105be752a05ad159b0e8d847286b9f2919b113d91b8" Feb 14 12:00:04 crc kubenswrapper[4904]: I0214 12:00:04.034926 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517840-ntbnl" Feb 14 12:00:04 crc kubenswrapper[4904]: I0214 12:00:04.436841 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7"] Feb 14 12:00:04 crc kubenswrapper[4904]: I0214 12:00:04.445772 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517795-5qrh7"] Feb 14 12:00:05 crc kubenswrapper[4904]: I0214 12:00:05.847597 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e31d6f5-239d-4418-b95d-dfe790798140" path="/var/lib/kubelet/pods/4e31d6f5-239d-4418-b95d-dfe790798140/volumes" Feb 14 12:00:08 crc kubenswrapper[4904]: I0214 12:00:08.836746 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 12:00:08 crc kubenswrapper[4904]: E0214 12:00:08.837281 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:00:16 crc kubenswrapper[4904]: I0214 12:00:16.129329 4904 scope.go:117] "RemoveContainer" containerID="0d5480a1226127a3e228870d0f6626435aa0a80ccd0821202d2f50a4b4e22b49" Feb 14 12:00:21 crc kubenswrapper[4904]: I0214 12:00:21.836917 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 12:00:21 crc kubenswrapper[4904]: E0214 12:00:21.837618 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:00:33 crc kubenswrapper[4904]: I0214 12:00:33.839237 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 12:00:33 crc kubenswrapper[4904]: E0214 12:00:33.840169 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:00:47 crc kubenswrapper[4904]: I0214 12:00:47.836271 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 12:00:47 crc kubenswrapper[4904]: E0214 12:00:47.836921 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:01:00 crc kubenswrapper[4904]: I0214 12:01:00.158792 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29517841-h7rq2"] Feb 14 12:01:00 crc kubenswrapper[4904]: E0214 12:01:00.160191 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16ca79af-7c49-4b29-96d5-87dca2fb8ead" containerName="collect-profiles" Feb 14 12:01:00 crc kubenswrapper[4904]: I0214 12:01:00.160213 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="16ca79af-7c49-4b29-96d5-87dca2fb8ead" containerName="collect-profiles" Feb 14 12:01:00 crc kubenswrapper[4904]: I0214 12:01:00.160423 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="16ca79af-7c49-4b29-96d5-87dca2fb8ead" containerName="collect-profiles" Feb 14 12:01:00 crc kubenswrapper[4904]: I0214 12:01:00.161276 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29517841-h7rq2" Feb 14 12:01:00 crc kubenswrapper[4904]: I0214 12:01:00.171539 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29517841-h7rq2"] Feb 14 12:01:00 crc kubenswrapper[4904]: I0214 12:01:00.241091 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtxxt\" (UniqueName: \"kubernetes.io/projected/629caf51-0db7-4bd8-80c2-59f45c9e09d2-kube-api-access-rtxxt\") pod \"keystone-cron-29517841-h7rq2\" (UID: \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\") " pod="openstack/keystone-cron-29517841-h7rq2" Feb 14 12:01:00 crc kubenswrapper[4904]: I0214 12:01:00.241142 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/629caf51-0db7-4bd8-80c2-59f45c9e09d2-fernet-keys\") pod \"keystone-cron-29517841-h7rq2\" (UID: \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\") " pod="openstack/keystone-cron-29517841-h7rq2" Feb 14 12:01:00 crc kubenswrapper[4904]: I0214 12:01:00.241220 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629caf51-0db7-4bd8-80c2-59f45c9e09d2-combined-ca-bundle\") pod \"keystone-cron-29517841-h7rq2\" (UID: \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\") " pod="openstack/keystone-cron-29517841-h7rq2" Feb 14 12:01:00 crc kubenswrapper[4904]: I0214 12:01:00.241283 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/629caf51-0db7-4bd8-80c2-59f45c9e09d2-config-data\") pod \"keystone-cron-29517841-h7rq2\" (UID: \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\") " pod="openstack/keystone-cron-29517841-h7rq2" Feb 14 12:01:00 crc kubenswrapper[4904]: I0214 12:01:00.342816 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtxxt\" (UniqueName: \"kubernetes.io/projected/629caf51-0db7-4bd8-80c2-59f45c9e09d2-kube-api-access-rtxxt\") pod \"keystone-cron-29517841-h7rq2\" (UID: \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\") " pod="openstack/keystone-cron-29517841-h7rq2" Feb 14 12:01:00 crc kubenswrapper[4904]: I0214 12:01:00.343107 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/629caf51-0db7-4bd8-80c2-59f45c9e09d2-fernet-keys\") pod \"keystone-cron-29517841-h7rq2\" (UID: \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\") " pod="openstack/keystone-cron-29517841-h7rq2" Feb 14 12:01:00 crc kubenswrapper[4904]: I0214 12:01:00.343153 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629caf51-0db7-4bd8-80c2-59f45c9e09d2-combined-ca-bundle\") pod \"keystone-cron-29517841-h7rq2\" (UID: \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\") " pod="openstack/keystone-cron-29517841-h7rq2" Feb 14 12:01:00 crc kubenswrapper[4904]: I0214 12:01:00.343211 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/629caf51-0db7-4bd8-80c2-59f45c9e09d2-config-data\") pod \"keystone-cron-29517841-h7rq2\" (UID: \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\") " pod="openstack/keystone-cron-29517841-h7rq2" Feb 14 12:01:00 crc kubenswrapper[4904]: I0214 12:01:00.348861 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629caf51-0db7-4bd8-80c2-59f45c9e09d2-combined-ca-bundle\") pod \"keystone-cron-29517841-h7rq2\" (UID: \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\") " pod="openstack/keystone-cron-29517841-h7rq2" Feb 14 12:01:00 crc kubenswrapper[4904]: I0214 12:01:00.349416 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/629caf51-0db7-4bd8-80c2-59f45c9e09d2-config-data\") pod \"keystone-cron-29517841-h7rq2\" (UID: \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\") " pod="openstack/keystone-cron-29517841-h7rq2" Feb 14 12:01:00 crc kubenswrapper[4904]: I0214 12:01:00.350186 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/629caf51-0db7-4bd8-80c2-59f45c9e09d2-fernet-keys\") pod \"keystone-cron-29517841-h7rq2\" (UID: \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\") " pod="openstack/keystone-cron-29517841-h7rq2" Feb 14 12:01:00 crc kubenswrapper[4904]: I0214 12:01:00.359552 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtxxt\" (UniqueName: \"kubernetes.io/projected/629caf51-0db7-4bd8-80c2-59f45c9e09d2-kube-api-access-rtxxt\") pod \"keystone-cron-29517841-h7rq2\" (UID: \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\") " pod="openstack/keystone-cron-29517841-h7rq2" Feb 14 12:01:00 crc kubenswrapper[4904]: I0214 12:01:00.484453 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29517841-h7rq2" Feb 14 12:01:00 crc kubenswrapper[4904]: I0214 12:01:00.930520 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29517841-h7rq2"] Feb 14 12:01:01 crc kubenswrapper[4904]: I0214 12:01:01.495774 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29517841-h7rq2" event={"ID":"629caf51-0db7-4bd8-80c2-59f45c9e09d2","Type":"ContainerStarted","Data":"4b10d79dc63689ac6cc5ca4cd92e5f20f167bfc6ce934de07f58850530b63ab7"} Feb 14 12:01:01 crc kubenswrapper[4904]: I0214 12:01:01.496143 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29517841-h7rq2" event={"ID":"629caf51-0db7-4bd8-80c2-59f45c9e09d2","Type":"ContainerStarted","Data":"be5459e0042f156535fc18b0583f7848739dc457db97b5c07a9fc2d3cae29826"} Feb 14 12:01:01 crc kubenswrapper[4904]: I0214 12:01:01.522487 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29517841-h7rq2" podStartSLOduration=1.52246819 podStartE2EDuration="1.52246819s" podCreationTimestamp="2026-02-14 12:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-14 12:01:01.511668253 +0000 UTC m=+3052.324432924" watchObservedRunningTime="2026-02-14 12:01:01.52246819 +0000 UTC m=+3052.335232861" Feb 14 12:01:01 crc kubenswrapper[4904]: I0214 12:01:01.840489 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 12:01:01 crc kubenswrapper[4904]: E0214 12:01:01.840794 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:01:04 crc kubenswrapper[4904]: I0214 12:01:04.520439 4904 generic.go:334] "Generic (PLEG): container finished" podID="629caf51-0db7-4bd8-80c2-59f45c9e09d2" containerID="4b10d79dc63689ac6cc5ca4cd92e5f20f167bfc6ce934de07f58850530b63ab7" exitCode=0 Feb 14 12:01:04 crc kubenswrapper[4904]: I0214 12:01:04.520528 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29517841-h7rq2" event={"ID":"629caf51-0db7-4bd8-80c2-59f45c9e09d2","Type":"ContainerDied","Data":"4b10d79dc63689ac6cc5ca4cd92e5f20f167bfc6ce934de07f58850530b63ab7"} Feb 14 12:01:05 crc kubenswrapper[4904]: I0214 12:01:05.829968 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29517841-h7rq2" Feb 14 12:01:05 crc kubenswrapper[4904]: I0214 12:01:05.861340 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtxxt\" (UniqueName: \"kubernetes.io/projected/629caf51-0db7-4bd8-80c2-59f45c9e09d2-kube-api-access-rtxxt\") pod \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\" (UID: \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\") " Feb 14 12:01:05 crc kubenswrapper[4904]: I0214 12:01:05.861381 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/629caf51-0db7-4bd8-80c2-59f45c9e09d2-fernet-keys\") pod \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\" (UID: \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\") " Feb 14 12:01:05 crc kubenswrapper[4904]: I0214 12:01:05.861403 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/629caf51-0db7-4bd8-80c2-59f45c9e09d2-config-data\") pod \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\" (UID: \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\") " Feb 14 12:01:05 crc kubenswrapper[4904]: I0214 12:01:05.861428 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629caf51-0db7-4bd8-80c2-59f45c9e09d2-combined-ca-bundle\") pod \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\" (UID: \"629caf51-0db7-4bd8-80c2-59f45c9e09d2\") " Feb 14 12:01:05 crc kubenswrapper[4904]: I0214 12:01:05.869805 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/629caf51-0db7-4bd8-80c2-59f45c9e09d2-kube-api-access-rtxxt" (OuterVolumeSpecName: "kube-api-access-rtxxt") pod "629caf51-0db7-4bd8-80c2-59f45c9e09d2" (UID: "629caf51-0db7-4bd8-80c2-59f45c9e09d2"). InnerVolumeSpecName "kube-api-access-rtxxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:01:05 crc kubenswrapper[4904]: I0214 12:01:05.871626 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/629caf51-0db7-4bd8-80c2-59f45c9e09d2-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "629caf51-0db7-4bd8-80c2-59f45c9e09d2" (UID: "629caf51-0db7-4bd8-80c2-59f45c9e09d2"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 12:01:05 crc kubenswrapper[4904]: I0214 12:01:05.891380 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/629caf51-0db7-4bd8-80c2-59f45c9e09d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "629caf51-0db7-4bd8-80c2-59f45c9e09d2" (UID: "629caf51-0db7-4bd8-80c2-59f45c9e09d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 12:01:05 crc kubenswrapper[4904]: I0214 12:01:05.919743 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/629caf51-0db7-4bd8-80c2-59f45c9e09d2-config-data" (OuterVolumeSpecName: "config-data") pod "629caf51-0db7-4bd8-80c2-59f45c9e09d2" (UID: "629caf51-0db7-4bd8-80c2-59f45c9e09d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 12:01:05 crc kubenswrapper[4904]: I0214 12:01:05.963004 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtxxt\" (UniqueName: \"kubernetes.io/projected/629caf51-0db7-4bd8-80c2-59f45c9e09d2-kube-api-access-rtxxt\") on node \"crc\" DevicePath \"\"" Feb 14 12:01:05 crc kubenswrapper[4904]: I0214 12:01:05.963051 4904 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/629caf51-0db7-4bd8-80c2-59f45c9e09d2-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 14 12:01:05 crc kubenswrapper[4904]: I0214 12:01:05.963064 4904 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/629caf51-0db7-4bd8-80c2-59f45c9e09d2-config-data\") on node \"crc\" DevicePath \"\"" Feb 14 12:01:05 crc kubenswrapper[4904]: I0214 12:01:05.963077 4904 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629caf51-0db7-4bd8-80c2-59f45c9e09d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 14 12:01:06 crc kubenswrapper[4904]: I0214 12:01:06.538308 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29517841-h7rq2" event={"ID":"629caf51-0db7-4bd8-80c2-59f45c9e09d2","Type":"ContainerDied","Data":"be5459e0042f156535fc18b0583f7848739dc457db97b5c07a9fc2d3cae29826"} Feb 14 12:01:06 crc kubenswrapper[4904]: I0214 12:01:06.538646 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be5459e0042f156535fc18b0583f7848739dc457db97b5c07a9fc2d3cae29826" Feb 14 12:01:06 crc kubenswrapper[4904]: I0214 12:01:06.538374 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29517841-h7rq2" Feb 14 12:01:15 crc kubenswrapper[4904]: I0214 12:01:15.856513 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 12:01:15 crc kubenswrapper[4904]: E0214 12:01:15.857303 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:01:27 crc kubenswrapper[4904]: I0214 12:01:27.836294 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 12:01:27 crc kubenswrapper[4904]: E0214 12:01:27.837083 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:01:38 crc kubenswrapper[4904]: I0214 12:01:38.836646 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 12:01:38 crc kubenswrapper[4904]: E0214 12:01:38.837459 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:01:46 crc kubenswrapper[4904]: I0214 12:01:46.709932 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-whbmz"] Feb 14 12:01:46 crc kubenswrapper[4904]: E0214 12:01:46.710919 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="629caf51-0db7-4bd8-80c2-59f45c9e09d2" containerName="keystone-cron" Feb 14 12:01:46 crc kubenswrapper[4904]: I0214 12:01:46.710935 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="629caf51-0db7-4bd8-80c2-59f45c9e09d2" containerName="keystone-cron" Feb 14 12:01:46 crc kubenswrapper[4904]: I0214 12:01:46.711175 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="629caf51-0db7-4bd8-80c2-59f45c9e09d2" containerName="keystone-cron" Feb 14 12:01:46 crc kubenswrapper[4904]: I0214 12:01:46.712864 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-whbmz" Feb 14 12:01:46 crc kubenswrapper[4904]: I0214 12:01:46.721976 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-whbmz"] Feb 14 12:01:46 crc kubenswrapper[4904]: I0214 12:01:46.782176 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76be7630-5a6a-4651-bb8b-5623da6169bb-catalog-content\") pod \"redhat-marketplace-whbmz\" (UID: \"76be7630-5a6a-4651-bb8b-5623da6169bb\") " pod="openshift-marketplace/redhat-marketplace-whbmz" Feb 14 12:01:46 crc kubenswrapper[4904]: I0214 12:01:46.782223 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76be7630-5a6a-4651-bb8b-5623da6169bb-utilities\") pod \"redhat-marketplace-whbmz\" (UID: \"76be7630-5a6a-4651-bb8b-5623da6169bb\") " pod="openshift-marketplace/redhat-marketplace-whbmz" Feb 14 12:01:46 crc kubenswrapper[4904]: I0214 12:01:46.782253 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8fbm\" (UniqueName: \"kubernetes.io/projected/76be7630-5a6a-4651-bb8b-5623da6169bb-kube-api-access-x8fbm\") pod \"redhat-marketplace-whbmz\" (UID: \"76be7630-5a6a-4651-bb8b-5623da6169bb\") " pod="openshift-marketplace/redhat-marketplace-whbmz" Feb 14 12:01:46 crc kubenswrapper[4904]: I0214 12:01:46.884266 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76be7630-5a6a-4651-bb8b-5623da6169bb-catalog-content\") pod \"redhat-marketplace-whbmz\" (UID: \"76be7630-5a6a-4651-bb8b-5623da6169bb\") " pod="openshift-marketplace/redhat-marketplace-whbmz" Feb 14 12:01:46 crc kubenswrapper[4904]: I0214 12:01:46.884332 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76be7630-5a6a-4651-bb8b-5623da6169bb-utilities\") pod \"redhat-marketplace-whbmz\" (UID: \"76be7630-5a6a-4651-bb8b-5623da6169bb\") " pod="openshift-marketplace/redhat-marketplace-whbmz" Feb 14 12:01:46 crc kubenswrapper[4904]: I0214 12:01:46.885000 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76be7630-5a6a-4651-bb8b-5623da6169bb-catalog-content\") pod \"redhat-marketplace-whbmz\" (UID: \"76be7630-5a6a-4651-bb8b-5623da6169bb\") " pod="openshift-marketplace/redhat-marketplace-whbmz" Feb 14 12:01:46 crc kubenswrapper[4904]: I0214 12:01:46.885044 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76be7630-5a6a-4651-bb8b-5623da6169bb-utilities\") pod \"redhat-marketplace-whbmz\" (UID: \"76be7630-5a6a-4651-bb8b-5623da6169bb\") " pod="openshift-marketplace/redhat-marketplace-whbmz" Feb 14 12:01:46 crc kubenswrapper[4904]: I0214 12:01:46.884363 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8fbm\" (UniqueName: \"kubernetes.io/projected/76be7630-5a6a-4651-bb8b-5623da6169bb-kube-api-access-x8fbm\") pod \"redhat-marketplace-whbmz\" (UID: \"76be7630-5a6a-4651-bb8b-5623da6169bb\") " pod="openshift-marketplace/redhat-marketplace-whbmz" Feb 14 12:01:46 crc kubenswrapper[4904]: I0214 12:01:46.903528 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8fbm\" (UniqueName: \"kubernetes.io/projected/76be7630-5a6a-4651-bb8b-5623da6169bb-kube-api-access-x8fbm\") pod \"redhat-marketplace-whbmz\" (UID: \"76be7630-5a6a-4651-bb8b-5623da6169bb\") " pod="openshift-marketplace/redhat-marketplace-whbmz" Feb 14 12:01:47 crc kubenswrapper[4904]: I0214 12:01:47.074268 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-whbmz" Feb 14 12:01:47 crc kubenswrapper[4904]: W0214 12:01:47.538602 4904 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76be7630_5a6a_4651_bb8b_5623da6169bb.slice/crio-db3d4abaf6193c9435673295692b16929cb0723a5d7712edc7e7ae33af800372 WatchSource:0}: Error finding container db3d4abaf6193c9435673295692b16929cb0723a5d7712edc7e7ae33af800372: Status 404 returned error can't find the container with id db3d4abaf6193c9435673295692b16929cb0723a5d7712edc7e7ae33af800372 Feb 14 12:01:47 crc kubenswrapper[4904]: I0214 12:01:47.549576 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-whbmz"] Feb 14 12:01:47 crc kubenswrapper[4904]: I0214 12:01:47.877579 4904 generic.go:334] "Generic (PLEG): container finished" podID="76be7630-5a6a-4651-bb8b-5623da6169bb" containerID="ef4f88698ed4502357bfc4e290e26f0604d70b4284e7846dea2d234990d4c652" exitCode=0 Feb 14 12:01:47 crc kubenswrapper[4904]: I0214 12:01:47.877621 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whbmz" event={"ID":"76be7630-5a6a-4651-bb8b-5623da6169bb","Type":"ContainerDied","Data":"ef4f88698ed4502357bfc4e290e26f0604d70b4284e7846dea2d234990d4c652"} Feb 14 12:01:47 crc kubenswrapper[4904]: I0214 12:01:47.877647 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whbmz" event={"ID":"76be7630-5a6a-4651-bb8b-5623da6169bb","Type":"ContainerStarted","Data":"db3d4abaf6193c9435673295692b16929cb0723a5d7712edc7e7ae33af800372"} Feb 14 12:01:47 crc kubenswrapper[4904]: I0214 12:01:47.881057 4904 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 14 12:01:48 crc kubenswrapper[4904]: I0214 12:01:48.886027 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whbmz" event={"ID":"76be7630-5a6a-4651-bb8b-5623da6169bb","Type":"ContainerStarted","Data":"6f289c159390c1b29fa8ebe47a52d1b13137dcdfc877d270ca0e6a80bba0acdb"} Feb 14 12:01:49 crc kubenswrapper[4904]: I0214 12:01:49.899611 4904 generic.go:334] "Generic (PLEG): container finished" podID="76be7630-5a6a-4651-bb8b-5623da6169bb" containerID="6f289c159390c1b29fa8ebe47a52d1b13137dcdfc877d270ca0e6a80bba0acdb" exitCode=0 Feb 14 12:01:49 crc kubenswrapper[4904]: I0214 12:01:49.899667 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whbmz" event={"ID":"76be7630-5a6a-4651-bb8b-5623da6169bb","Type":"ContainerDied","Data":"6f289c159390c1b29fa8ebe47a52d1b13137dcdfc877d270ca0e6a80bba0acdb"} Feb 14 12:01:50 crc kubenswrapper[4904]: I0214 12:01:50.910736 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whbmz" event={"ID":"76be7630-5a6a-4651-bb8b-5623da6169bb","Type":"ContainerStarted","Data":"1136502f7025939c18bb68d5ce81fc163645102f6eb906742b54642c64256ce0"} Feb 14 12:01:50 crc kubenswrapper[4904]: I0214 12:01:50.929981 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-whbmz" podStartSLOduration=2.497455863 podStartE2EDuration="4.929963335s" podCreationTimestamp="2026-02-14 12:01:46 +0000 UTC" firstStartedPulling="2026-02-14 12:01:47.880818963 +0000 UTC m=+3098.693583624" lastFinishedPulling="2026-02-14 12:01:50.313326435 +0000 UTC m=+3101.126091096" observedRunningTime="2026-02-14 12:01:50.929655137 +0000 UTC m=+3101.742419798" watchObservedRunningTime="2026-02-14 12:01:50.929963335 +0000 UTC m=+3101.742727996" Feb 14 12:01:51 crc kubenswrapper[4904]: I0214 12:01:51.836518 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 12:01:51 crc kubenswrapper[4904]: E0214 12:01:51.837033 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:01:57 crc kubenswrapper[4904]: I0214 12:01:57.074560 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-whbmz" Feb 14 12:01:57 crc kubenswrapper[4904]: I0214 12:01:57.075182 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-whbmz" Feb 14 12:01:57 crc kubenswrapper[4904]: I0214 12:01:57.131826 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-whbmz" Feb 14 12:01:58 crc kubenswrapper[4904]: I0214 12:01:58.034113 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-whbmz" Feb 14 12:01:58 crc kubenswrapper[4904]: I0214 12:01:58.077112 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-whbmz"] Feb 14 12:01:59 crc kubenswrapper[4904]: I0214 12:01:59.996272 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-whbmz" podUID="76be7630-5a6a-4651-bb8b-5623da6169bb" containerName="registry-server" containerID="cri-o://1136502f7025939c18bb68d5ce81fc163645102f6eb906742b54642c64256ce0" gracePeriod=2 Feb 14 12:02:00 crc kubenswrapper[4904]: I0214 12:02:00.414925 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-whbmz" Feb 14 12:02:00 crc kubenswrapper[4904]: I0214 12:02:00.426400 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76be7630-5a6a-4651-bb8b-5623da6169bb-utilities\") pod \"76be7630-5a6a-4651-bb8b-5623da6169bb\" (UID: \"76be7630-5a6a-4651-bb8b-5623da6169bb\") " Feb 14 12:02:00 crc kubenswrapper[4904]: I0214 12:02:00.426469 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8fbm\" (UniqueName: \"kubernetes.io/projected/76be7630-5a6a-4651-bb8b-5623da6169bb-kube-api-access-x8fbm\") pod \"76be7630-5a6a-4651-bb8b-5623da6169bb\" (UID: \"76be7630-5a6a-4651-bb8b-5623da6169bb\") " Feb 14 12:02:00 crc kubenswrapper[4904]: I0214 12:02:00.426593 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76be7630-5a6a-4651-bb8b-5623da6169bb-catalog-content\") pod \"76be7630-5a6a-4651-bb8b-5623da6169bb\" (UID: \"76be7630-5a6a-4651-bb8b-5623da6169bb\") " Feb 14 12:02:00 crc kubenswrapper[4904]: I0214 12:02:00.428156 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76be7630-5a6a-4651-bb8b-5623da6169bb-utilities" (OuterVolumeSpecName: "utilities") pod "76be7630-5a6a-4651-bb8b-5623da6169bb" (UID: "76be7630-5a6a-4651-bb8b-5623da6169bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:02:00 crc kubenswrapper[4904]: I0214 12:02:00.442069 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76be7630-5a6a-4651-bb8b-5623da6169bb-kube-api-access-x8fbm" (OuterVolumeSpecName: "kube-api-access-x8fbm") pod "76be7630-5a6a-4651-bb8b-5623da6169bb" (UID: "76be7630-5a6a-4651-bb8b-5623da6169bb"). InnerVolumeSpecName "kube-api-access-x8fbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:02:00 crc kubenswrapper[4904]: I0214 12:02:00.465278 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76be7630-5a6a-4651-bb8b-5623da6169bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76be7630-5a6a-4651-bb8b-5623da6169bb" (UID: "76be7630-5a6a-4651-bb8b-5623da6169bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:02:00 crc kubenswrapper[4904]: I0214 12:02:00.528573 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76be7630-5a6a-4651-bb8b-5623da6169bb-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 12:02:00 crc kubenswrapper[4904]: I0214 12:02:00.528809 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8fbm\" (UniqueName: \"kubernetes.io/projected/76be7630-5a6a-4651-bb8b-5623da6169bb-kube-api-access-x8fbm\") on node \"crc\" DevicePath \"\"" Feb 14 12:02:00 crc kubenswrapper[4904]: I0214 12:02:00.528843 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76be7630-5a6a-4651-bb8b-5623da6169bb-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 12:02:01 crc kubenswrapper[4904]: I0214 12:02:01.007403 4904 generic.go:334] "Generic (PLEG): container finished" podID="76be7630-5a6a-4651-bb8b-5623da6169bb" containerID="1136502f7025939c18bb68d5ce81fc163645102f6eb906742b54642c64256ce0" exitCode=0 Feb 14 12:02:01 crc kubenswrapper[4904]: I0214 12:02:01.007446 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whbmz" event={"ID":"76be7630-5a6a-4651-bb8b-5623da6169bb","Type":"ContainerDied","Data":"1136502f7025939c18bb68d5ce81fc163645102f6eb906742b54642c64256ce0"} Feb 14 12:02:01 crc kubenswrapper[4904]: I0214 12:02:01.007466 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-whbmz" Feb 14 12:02:01 crc kubenswrapper[4904]: I0214 12:02:01.007476 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whbmz" event={"ID":"76be7630-5a6a-4651-bb8b-5623da6169bb","Type":"ContainerDied","Data":"db3d4abaf6193c9435673295692b16929cb0723a5d7712edc7e7ae33af800372"} Feb 14 12:02:01 crc kubenswrapper[4904]: I0214 12:02:01.007494 4904 scope.go:117] "RemoveContainer" containerID="1136502f7025939c18bb68d5ce81fc163645102f6eb906742b54642c64256ce0" Feb 14 12:02:01 crc kubenswrapper[4904]: I0214 12:02:01.037358 4904 scope.go:117] "RemoveContainer" containerID="6f289c159390c1b29fa8ebe47a52d1b13137dcdfc877d270ca0e6a80bba0acdb" Feb 14 12:02:01 crc kubenswrapper[4904]: I0214 12:02:01.055166 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-whbmz"] Feb 14 12:02:01 crc kubenswrapper[4904]: I0214 12:02:01.057888 4904 scope.go:117] "RemoveContainer" containerID="ef4f88698ed4502357bfc4e290e26f0604d70b4284e7846dea2d234990d4c652" Feb 14 12:02:01 crc kubenswrapper[4904]: I0214 12:02:01.064765 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-whbmz"] Feb 14 12:02:01 crc kubenswrapper[4904]: I0214 12:02:01.102084 4904 scope.go:117] "RemoveContainer" containerID="1136502f7025939c18bb68d5ce81fc163645102f6eb906742b54642c64256ce0" Feb 14 12:02:01 crc kubenswrapper[4904]: E0214 12:02:01.102544 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1136502f7025939c18bb68d5ce81fc163645102f6eb906742b54642c64256ce0\": container with ID starting with 1136502f7025939c18bb68d5ce81fc163645102f6eb906742b54642c64256ce0 not found: ID does not exist" containerID="1136502f7025939c18bb68d5ce81fc163645102f6eb906742b54642c64256ce0" Feb 14 12:02:01 crc kubenswrapper[4904]: I0214 12:02:01.102590 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1136502f7025939c18bb68d5ce81fc163645102f6eb906742b54642c64256ce0"} err="failed to get container status \"1136502f7025939c18bb68d5ce81fc163645102f6eb906742b54642c64256ce0\": rpc error: code = NotFound desc = could not find container \"1136502f7025939c18bb68d5ce81fc163645102f6eb906742b54642c64256ce0\": container with ID starting with 1136502f7025939c18bb68d5ce81fc163645102f6eb906742b54642c64256ce0 not found: ID does not exist" Feb 14 12:02:01 crc kubenswrapper[4904]: I0214 12:02:01.102618 4904 scope.go:117] "RemoveContainer" containerID="6f289c159390c1b29fa8ebe47a52d1b13137dcdfc877d270ca0e6a80bba0acdb" Feb 14 12:02:01 crc kubenswrapper[4904]: E0214 12:02:01.102935 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f289c159390c1b29fa8ebe47a52d1b13137dcdfc877d270ca0e6a80bba0acdb\": container with ID starting with 6f289c159390c1b29fa8ebe47a52d1b13137dcdfc877d270ca0e6a80bba0acdb not found: ID does not exist" containerID="6f289c159390c1b29fa8ebe47a52d1b13137dcdfc877d270ca0e6a80bba0acdb" Feb 14 12:02:01 crc kubenswrapper[4904]: I0214 12:02:01.102953 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f289c159390c1b29fa8ebe47a52d1b13137dcdfc877d270ca0e6a80bba0acdb"} err="failed to get container status \"6f289c159390c1b29fa8ebe47a52d1b13137dcdfc877d270ca0e6a80bba0acdb\": rpc error: code = NotFound desc = could not find container \"6f289c159390c1b29fa8ebe47a52d1b13137dcdfc877d270ca0e6a80bba0acdb\": container with ID starting with 6f289c159390c1b29fa8ebe47a52d1b13137dcdfc877d270ca0e6a80bba0acdb not found: ID does not exist" Feb 14 12:02:01 crc kubenswrapper[4904]: I0214 12:02:01.102966 4904 scope.go:117] "RemoveContainer" containerID="ef4f88698ed4502357bfc4e290e26f0604d70b4284e7846dea2d234990d4c652" Feb 14 12:02:01 crc kubenswrapper[4904]: E0214 12:02:01.103316 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef4f88698ed4502357bfc4e290e26f0604d70b4284e7846dea2d234990d4c652\": container with ID starting with ef4f88698ed4502357bfc4e290e26f0604d70b4284e7846dea2d234990d4c652 not found: ID does not exist" containerID="ef4f88698ed4502357bfc4e290e26f0604d70b4284e7846dea2d234990d4c652" Feb 14 12:02:01 crc kubenswrapper[4904]: I0214 12:02:01.103366 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef4f88698ed4502357bfc4e290e26f0604d70b4284e7846dea2d234990d4c652"} err="failed to get container status \"ef4f88698ed4502357bfc4e290e26f0604d70b4284e7846dea2d234990d4c652\": rpc error: code = NotFound desc = could not find container \"ef4f88698ed4502357bfc4e290e26f0604d70b4284e7846dea2d234990d4c652\": container with ID starting with ef4f88698ed4502357bfc4e290e26f0604d70b4284e7846dea2d234990d4c652 not found: ID does not exist" Feb 14 12:02:01 crc kubenswrapper[4904]: I0214 12:02:01.846004 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76be7630-5a6a-4651-bb8b-5623da6169bb" path="/var/lib/kubelet/pods/76be7630-5a6a-4651-bb8b-5623da6169bb/volumes" Feb 14 12:02:05 crc kubenswrapper[4904]: I0214 12:02:05.836886 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 12:02:05 crc kubenswrapper[4904]: E0214 12:02:05.838023 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:02:18 crc kubenswrapper[4904]: I0214 12:02:18.836648 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 12:02:18 crc kubenswrapper[4904]: E0214 12:02:18.837424 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:02:23 crc kubenswrapper[4904]: I0214 12:02:23.509395 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-npdct"] Feb 14 12:02:23 crc kubenswrapper[4904]: E0214 12:02:23.510297 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76be7630-5a6a-4651-bb8b-5623da6169bb" containerName="extract-content" Feb 14 12:02:23 crc kubenswrapper[4904]: I0214 12:02:23.510311 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="76be7630-5a6a-4651-bb8b-5623da6169bb" containerName="extract-content" Feb 14 12:02:23 crc kubenswrapper[4904]: E0214 12:02:23.510356 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76be7630-5a6a-4651-bb8b-5623da6169bb" containerName="registry-server" Feb 14 12:02:23 crc kubenswrapper[4904]: I0214 12:02:23.510362 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="76be7630-5a6a-4651-bb8b-5623da6169bb" containerName="registry-server" Feb 14 12:02:23 crc kubenswrapper[4904]: E0214 12:02:23.510373 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76be7630-5a6a-4651-bb8b-5623da6169bb" containerName="extract-utilities" Feb 14 12:02:23 crc kubenswrapper[4904]: I0214 12:02:23.510379 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="76be7630-5a6a-4651-bb8b-5623da6169bb" containerName="extract-utilities" Feb 14 12:02:23 crc kubenswrapper[4904]: I0214 12:02:23.510542 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="76be7630-5a6a-4651-bb8b-5623da6169bb" containerName="registry-server" Feb 14 12:02:23 crc kubenswrapper[4904]: I0214 12:02:23.511846 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-npdct" Feb 14 12:02:23 crc kubenswrapper[4904]: I0214 12:02:23.525386 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-npdct"] Feb 14 12:02:23 crc kubenswrapper[4904]: I0214 12:02:23.557429 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b970167-fc67-4487-bb8e-2ca2b98bdfea-utilities\") pod \"community-operators-npdct\" (UID: \"7b970167-fc67-4487-bb8e-2ca2b98bdfea\") " pod="openshift-marketplace/community-operators-npdct" Feb 14 12:02:23 crc kubenswrapper[4904]: I0214 12:02:23.557553 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b970167-fc67-4487-bb8e-2ca2b98bdfea-catalog-content\") pod \"community-operators-npdct\" (UID: \"7b970167-fc67-4487-bb8e-2ca2b98bdfea\") " pod="openshift-marketplace/community-operators-npdct" Feb 14 12:02:23 crc kubenswrapper[4904]: I0214 12:02:23.557586 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dh8g\" (UniqueName: \"kubernetes.io/projected/7b970167-fc67-4487-bb8e-2ca2b98bdfea-kube-api-access-2dh8g\") pod \"community-operators-npdct\" (UID: \"7b970167-fc67-4487-bb8e-2ca2b98bdfea\") " pod="openshift-marketplace/community-operators-npdct" Feb 14 12:02:23 crc kubenswrapper[4904]: I0214 12:02:23.658220 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b970167-fc67-4487-bb8e-2ca2b98bdfea-utilities\") pod \"community-operators-npdct\" (UID: \"7b970167-fc67-4487-bb8e-2ca2b98bdfea\") " pod="openshift-marketplace/community-operators-npdct" Feb 14 12:02:23 crc kubenswrapper[4904]: I0214 12:02:23.658363 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b970167-fc67-4487-bb8e-2ca2b98bdfea-catalog-content\") pod \"community-operators-npdct\" (UID: \"7b970167-fc67-4487-bb8e-2ca2b98bdfea\") " pod="openshift-marketplace/community-operators-npdct" Feb 14 12:02:23 crc kubenswrapper[4904]: I0214 12:02:23.658394 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dh8g\" (UniqueName: \"kubernetes.io/projected/7b970167-fc67-4487-bb8e-2ca2b98bdfea-kube-api-access-2dh8g\") pod \"community-operators-npdct\" (UID: \"7b970167-fc67-4487-bb8e-2ca2b98bdfea\") " pod="openshift-marketplace/community-operators-npdct" Feb 14 12:02:23 crc kubenswrapper[4904]: I0214 12:02:23.658704 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b970167-fc67-4487-bb8e-2ca2b98bdfea-utilities\") pod \"community-operators-npdct\" (UID: \"7b970167-fc67-4487-bb8e-2ca2b98bdfea\") " pod="openshift-marketplace/community-operators-npdct" Feb 14 12:02:23 crc kubenswrapper[4904]: I0214 12:02:23.659010 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b970167-fc67-4487-bb8e-2ca2b98bdfea-catalog-content\") pod \"community-operators-npdct\" (UID: \"7b970167-fc67-4487-bb8e-2ca2b98bdfea\") " pod="openshift-marketplace/community-operators-npdct" Feb 14 12:02:23 crc kubenswrapper[4904]: I0214 12:02:23.680784 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dh8g\" (UniqueName: \"kubernetes.io/projected/7b970167-fc67-4487-bb8e-2ca2b98bdfea-kube-api-access-2dh8g\") pod \"community-operators-npdct\" (UID: \"7b970167-fc67-4487-bb8e-2ca2b98bdfea\") " pod="openshift-marketplace/community-operators-npdct" Feb 14 12:02:23 crc kubenswrapper[4904]: I0214 12:02:23.831179 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-npdct" Feb 14 12:02:24 crc kubenswrapper[4904]: I0214 12:02:24.350654 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-npdct"] Feb 14 12:02:25 crc kubenswrapper[4904]: I0214 12:02:25.207513 4904 generic.go:334] "Generic (PLEG): container finished" podID="7b970167-fc67-4487-bb8e-2ca2b98bdfea" containerID="501f794dac868eea863e94d90488719ef35d382682c7e23ecd2999535372c914" exitCode=0 Feb 14 12:02:25 crc kubenswrapper[4904]: I0214 12:02:25.207567 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-npdct" event={"ID":"7b970167-fc67-4487-bb8e-2ca2b98bdfea","Type":"ContainerDied","Data":"501f794dac868eea863e94d90488719ef35d382682c7e23ecd2999535372c914"} Feb 14 12:02:25 crc kubenswrapper[4904]: I0214 12:02:25.207617 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-npdct" event={"ID":"7b970167-fc67-4487-bb8e-2ca2b98bdfea","Type":"ContainerStarted","Data":"bd118fc82217786d42bc502756d57ecdee36b18a118ceabaa4582cd95d6b8c84"} Feb 14 12:02:26 crc kubenswrapper[4904]: I0214 12:02:26.219030 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-npdct" event={"ID":"7b970167-fc67-4487-bb8e-2ca2b98bdfea","Type":"ContainerStarted","Data":"1dd18cee83db8c4d68825d09efa2cff456d9c59b16cfc4ca7e1bfcb708726fc0"} Feb 14 12:02:27 crc kubenswrapper[4904]: I0214 12:02:27.231103 4904 generic.go:334] "Generic (PLEG): container finished" podID="7b970167-fc67-4487-bb8e-2ca2b98bdfea" containerID="1dd18cee83db8c4d68825d09efa2cff456d9c59b16cfc4ca7e1bfcb708726fc0" exitCode=0 Feb 14 12:02:27 crc kubenswrapper[4904]: I0214 12:02:27.231218 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-npdct" event={"ID":"7b970167-fc67-4487-bb8e-2ca2b98bdfea","Type":"ContainerDied","Data":"1dd18cee83db8c4d68825d09efa2cff456d9c59b16cfc4ca7e1bfcb708726fc0"} Feb 14 12:02:28 crc kubenswrapper[4904]: I0214 12:02:28.240614 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-npdct" event={"ID":"7b970167-fc67-4487-bb8e-2ca2b98bdfea","Type":"ContainerStarted","Data":"fb2a2a0b6fda03e528709681722e1f79a02b9d581aab99b7b9d047001edf260f"} Feb 14 12:02:28 crc kubenswrapper[4904]: I0214 12:02:28.265383 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-npdct" podStartSLOduration=2.886088063 podStartE2EDuration="5.265362773s" podCreationTimestamp="2026-02-14 12:02:23 +0000 UTC" firstStartedPulling="2026-02-14 12:02:25.209147588 +0000 UTC m=+3136.021912249" lastFinishedPulling="2026-02-14 12:02:27.588422298 +0000 UTC m=+3138.401186959" observedRunningTime="2026-02-14 12:02:28.263529262 +0000 UTC m=+3139.076293923" watchObservedRunningTime="2026-02-14 12:02:28.265362773 +0000 UTC m=+3139.078127444" Feb 14 12:02:32 crc kubenswrapper[4904]: I0214 12:02:32.835878 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 12:02:32 crc kubenswrapper[4904]: E0214 12:02:32.836799 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:02:33 crc kubenswrapper[4904]: I0214 12:02:33.831698 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-npdct" Feb 14 12:02:33 crc kubenswrapper[4904]: I0214 12:02:33.831746 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-npdct" Feb 14 12:02:33 crc kubenswrapper[4904]: I0214 12:02:33.878758 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-npdct" Feb 14 12:02:34 crc kubenswrapper[4904]: I0214 12:02:34.327632 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-npdct" Feb 14 12:02:34 crc kubenswrapper[4904]: I0214 12:02:34.380259 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-npdct"] Feb 14 12:02:36 crc kubenswrapper[4904]: I0214 12:02:36.295729 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-npdct" podUID="7b970167-fc67-4487-bb8e-2ca2b98bdfea" containerName="registry-server" containerID="cri-o://fb2a2a0b6fda03e528709681722e1f79a02b9d581aab99b7b9d047001edf260f" gracePeriod=2 Feb 14 12:02:36 crc kubenswrapper[4904]: I0214 12:02:36.699326 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-npdct" Feb 14 12:02:36 crc kubenswrapper[4904]: I0214 12:02:36.803252 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dh8g\" (UniqueName: \"kubernetes.io/projected/7b970167-fc67-4487-bb8e-2ca2b98bdfea-kube-api-access-2dh8g\") pod \"7b970167-fc67-4487-bb8e-2ca2b98bdfea\" (UID: \"7b970167-fc67-4487-bb8e-2ca2b98bdfea\") " Feb 14 12:02:36 crc kubenswrapper[4904]: I0214 12:02:36.803360 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b970167-fc67-4487-bb8e-2ca2b98bdfea-utilities\") pod \"7b970167-fc67-4487-bb8e-2ca2b98bdfea\" (UID: \"7b970167-fc67-4487-bb8e-2ca2b98bdfea\") " Feb 14 12:02:36 crc kubenswrapper[4904]: I0214 12:02:36.803468 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b970167-fc67-4487-bb8e-2ca2b98bdfea-catalog-content\") pod \"7b970167-fc67-4487-bb8e-2ca2b98bdfea\" (UID: \"7b970167-fc67-4487-bb8e-2ca2b98bdfea\") " Feb 14 12:02:36 crc kubenswrapper[4904]: I0214 12:02:36.805448 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b970167-fc67-4487-bb8e-2ca2b98bdfea-utilities" (OuterVolumeSpecName: "utilities") pod "7b970167-fc67-4487-bb8e-2ca2b98bdfea" (UID: "7b970167-fc67-4487-bb8e-2ca2b98bdfea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:02:36 crc kubenswrapper[4904]: I0214 12:02:36.818143 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b970167-fc67-4487-bb8e-2ca2b98bdfea-kube-api-access-2dh8g" (OuterVolumeSpecName: "kube-api-access-2dh8g") pod "7b970167-fc67-4487-bb8e-2ca2b98bdfea" (UID: "7b970167-fc67-4487-bb8e-2ca2b98bdfea"). InnerVolumeSpecName "kube-api-access-2dh8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:02:36 crc kubenswrapper[4904]: I0214 12:02:36.854533 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b970167-fc67-4487-bb8e-2ca2b98bdfea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b970167-fc67-4487-bb8e-2ca2b98bdfea" (UID: "7b970167-fc67-4487-bb8e-2ca2b98bdfea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:02:36 crc kubenswrapper[4904]: I0214 12:02:36.905416 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dh8g\" (UniqueName: \"kubernetes.io/projected/7b970167-fc67-4487-bb8e-2ca2b98bdfea-kube-api-access-2dh8g\") on node \"crc\" DevicePath \"\"" Feb 14 12:02:36 crc kubenswrapper[4904]: I0214 12:02:36.905630 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b970167-fc67-4487-bb8e-2ca2b98bdfea-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 12:02:36 crc kubenswrapper[4904]: I0214 12:02:36.905689 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b970167-fc67-4487-bb8e-2ca2b98bdfea-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 12:02:37 crc kubenswrapper[4904]: I0214 12:02:37.310582 4904 generic.go:334] "Generic (PLEG): container finished" podID="7b970167-fc67-4487-bb8e-2ca2b98bdfea" containerID="fb2a2a0b6fda03e528709681722e1f79a02b9d581aab99b7b9d047001edf260f" exitCode=0 Feb 14 12:02:37 crc kubenswrapper[4904]: I0214 12:02:37.310642 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-npdct" event={"ID":"7b970167-fc67-4487-bb8e-2ca2b98bdfea","Type":"ContainerDied","Data":"fb2a2a0b6fda03e528709681722e1f79a02b9d581aab99b7b9d047001edf260f"} Feb 14 12:02:37 crc kubenswrapper[4904]: I0214 12:02:37.310933 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-npdct" event={"ID":"7b970167-fc67-4487-bb8e-2ca2b98bdfea","Type":"ContainerDied","Data":"bd118fc82217786d42bc502756d57ecdee36b18a118ceabaa4582cd95d6b8c84"} Feb 14 12:02:37 crc kubenswrapper[4904]: I0214 12:02:37.310955 4904 scope.go:117] "RemoveContainer" containerID="fb2a2a0b6fda03e528709681722e1f79a02b9d581aab99b7b9d047001edf260f" Feb 14 12:02:37 crc kubenswrapper[4904]: I0214 12:02:37.310661 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-npdct" Feb 14 12:02:37 crc kubenswrapper[4904]: I0214 12:02:37.344580 4904 scope.go:117] "RemoveContainer" containerID="1dd18cee83db8c4d68825d09efa2cff456d9c59b16cfc4ca7e1bfcb708726fc0" Feb 14 12:02:37 crc kubenswrapper[4904]: I0214 12:02:37.351346 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-npdct"] Feb 14 12:02:37 crc kubenswrapper[4904]: I0214 12:02:37.362759 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-npdct"] Feb 14 12:02:37 crc kubenswrapper[4904]: I0214 12:02:37.371112 4904 scope.go:117] "RemoveContainer" containerID="501f794dac868eea863e94d90488719ef35d382682c7e23ecd2999535372c914" Feb 14 12:02:37 crc kubenswrapper[4904]: I0214 12:02:37.411457 4904 scope.go:117] "RemoveContainer" containerID="fb2a2a0b6fda03e528709681722e1f79a02b9d581aab99b7b9d047001edf260f" Feb 14 12:02:37 crc kubenswrapper[4904]: E0214 12:02:37.412030 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb2a2a0b6fda03e528709681722e1f79a02b9d581aab99b7b9d047001edf260f\": container with ID starting with fb2a2a0b6fda03e528709681722e1f79a02b9d581aab99b7b9d047001edf260f not found: ID does not exist" containerID="fb2a2a0b6fda03e528709681722e1f79a02b9d581aab99b7b9d047001edf260f" Feb 14 12:02:37 crc kubenswrapper[4904]: I0214 12:02:37.412132 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb2a2a0b6fda03e528709681722e1f79a02b9d581aab99b7b9d047001edf260f"} err="failed to get container status \"fb2a2a0b6fda03e528709681722e1f79a02b9d581aab99b7b9d047001edf260f\": rpc error: code = NotFound desc = could not find container \"fb2a2a0b6fda03e528709681722e1f79a02b9d581aab99b7b9d047001edf260f\": container with ID starting with fb2a2a0b6fda03e528709681722e1f79a02b9d581aab99b7b9d047001edf260f not found: ID does not exist" Feb 14 12:02:37 crc kubenswrapper[4904]: I0214 12:02:37.412206 4904 scope.go:117] "RemoveContainer" containerID="1dd18cee83db8c4d68825d09efa2cff456d9c59b16cfc4ca7e1bfcb708726fc0" Feb 14 12:02:37 crc kubenswrapper[4904]: E0214 12:02:37.412621 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dd18cee83db8c4d68825d09efa2cff456d9c59b16cfc4ca7e1bfcb708726fc0\": container with ID starting with 1dd18cee83db8c4d68825d09efa2cff456d9c59b16cfc4ca7e1bfcb708726fc0 not found: ID does not exist" containerID="1dd18cee83db8c4d68825d09efa2cff456d9c59b16cfc4ca7e1bfcb708726fc0" Feb 14 12:02:37 crc kubenswrapper[4904]: I0214 12:02:37.412662 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dd18cee83db8c4d68825d09efa2cff456d9c59b16cfc4ca7e1bfcb708726fc0"} err="failed to get container status \"1dd18cee83db8c4d68825d09efa2cff456d9c59b16cfc4ca7e1bfcb708726fc0\": rpc error: code = NotFound desc = could not find container \"1dd18cee83db8c4d68825d09efa2cff456d9c59b16cfc4ca7e1bfcb708726fc0\": container with ID starting with 1dd18cee83db8c4d68825d09efa2cff456d9c59b16cfc4ca7e1bfcb708726fc0 not found: ID does not exist" Feb 14 12:02:37 crc kubenswrapper[4904]: I0214 12:02:37.412695 4904 scope.go:117] "RemoveContainer" containerID="501f794dac868eea863e94d90488719ef35d382682c7e23ecd2999535372c914" Feb 14 12:02:37 crc kubenswrapper[4904]: E0214 12:02:37.412969 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"501f794dac868eea863e94d90488719ef35d382682c7e23ecd2999535372c914\": container with ID starting with 501f794dac868eea863e94d90488719ef35d382682c7e23ecd2999535372c914 not found: ID does not exist" containerID="501f794dac868eea863e94d90488719ef35d382682c7e23ecd2999535372c914" Feb 14 12:02:37 crc kubenswrapper[4904]: I0214 12:02:37.413065 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"501f794dac868eea863e94d90488719ef35d382682c7e23ecd2999535372c914"} err="failed to get container status \"501f794dac868eea863e94d90488719ef35d382682c7e23ecd2999535372c914\": rpc error: code = NotFound desc = could not find container \"501f794dac868eea863e94d90488719ef35d382682c7e23ecd2999535372c914\": container with ID starting with 501f794dac868eea863e94d90488719ef35d382682c7e23ecd2999535372c914 not found: ID does not exist" Feb 14 12:02:37 crc kubenswrapper[4904]: I0214 12:02:37.846247 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b970167-fc67-4487-bb8e-2ca2b98bdfea" path="/var/lib/kubelet/pods/7b970167-fc67-4487-bb8e-2ca2b98bdfea/volumes" Feb 14 12:02:46 crc kubenswrapper[4904]: I0214 12:02:46.836422 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 12:02:46 crc kubenswrapper[4904]: E0214 12:02:46.837205 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:02:58 crc kubenswrapper[4904]: I0214 12:02:58.837326 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 12:02:58 crc kubenswrapper[4904]: E0214 12:02:58.838058 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:03:13 crc kubenswrapper[4904]: I0214 12:03:13.836934 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 12:03:13 crc kubenswrapper[4904]: E0214 12:03:13.839039 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:03:19 crc kubenswrapper[4904]: I0214 12:03:19.089739 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q6rh2"] Feb 14 12:03:19 crc kubenswrapper[4904]: E0214 12:03:19.092331 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b970167-fc67-4487-bb8e-2ca2b98bdfea" containerName="registry-server" Feb 14 12:03:19 crc kubenswrapper[4904]: I0214 12:03:19.092348 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b970167-fc67-4487-bb8e-2ca2b98bdfea" containerName="registry-server" Feb 14 12:03:19 crc kubenswrapper[4904]: E0214 12:03:19.092365 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b970167-fc67-4487-bb8e-2ca2b98bdfea" containerName="extract-content" Feb 14 12:03:19 crc kubenswrapper[4904]: I0214 12:03:19.092371 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b970167-fc67-4487-bb8e-2ca2b98bdfea" containerName="extract-content" Feb 14 12:03:19 crc kubenswrapper[4904]: E0214 12:03:19.092394 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b970167-fc67-4487-bb8e-2ca2b98bdfea" containerName="extract-utilities" Feb 14 12:03:19 crc kubenswrapper[4904]: I0214 12:03:19.092400 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b970167-fc67-4487-bb8e-2ca2b98bdfea" containerName="extract-utilities" Feb 14 12:03:19 crc kubenswrapper[4904]: I0214 12:03:19.092559 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b970167-fc67-4487-bb8e-2ca2b98bdfea" containerName="registry-server" Feb 14 12:03:19 crc kubenswrapper[4904]: I0214 12:03:19.093868 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6rh2" Feb 14 12:03:19 crc kubenswrapper[4904]: I0214 12:03:19.120888 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q6rh2"] Feb 14 12:03:19 crc kubenswrapper[4904]: I0214 12:03:19.210341 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c3d261f-9b09-48b3-b47a-fcbc4985b47f-catalog-content\") pod \"certified-operators-q6rh2\" (UID: \"2c3d261f-9b09-48b3-b47a-fcbc4985b47f\") " pod="openshift-marketplace/certified-operators-q6rh2" Feb 14 12:03:19 crc kubenswrapper[4904]: I0214 12:03:19.210435 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpj79\" (UniqueName: \"kubernetes.io/projected/2c3d261f-9b09-48b3-b47a-fcbc4985b47f-kube-api-access-wpj79\") pod \"certified-operators-q6rh2\" (UID: \"2c3d261f-9b09-48b3-b47a-fcbc4985b47f\") " pod="openshift-marketplace/certified-operators-q6rh2" Feb 14 12:03:19 crc kubenswrapper[4904]: I0214 12:03:19.210458 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c3d261f-9b09-48b3-b47a-fcbc4985b47f-utilities\") pod \"certified-operators-q6rh2\" (UID: \"2c3d261f-9b09-48b3-b47a-fcbc4985b47f\") " pod="openshift-marketplace/certified-operators-q6rh2" Feb 14 12:03:19 crc kubenswrapper[4904]: I0214 12:03:19.313801 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c3d261f-9b09-48b3-b47a-fcbc4985b47f-catalog-content\") pod \"certified-operators-q6rh2\" (UID: \"2c3d261f-9b09-48b3-b47a-fcbc4985b47f\") " pod="openshift-marketplace/certified-operators-q6rh2" Feb 14 12:03:19 crc kubenswrapper[4904]: I0214 12:03:19.313990 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpj79\" (UniqueName: \"kubernetes.io/projected/2c3d261f-9b09-48b3-b47a-fcbc4985b47f-kube-api-access-wpj79\") pod \"certified-operators-q6rh2\" (UID: \"2c3d261f-9b09-48b3-b47a-fcbc4985b47f\") " pod="openshift-marketplace/certified-operators-q6rh2" Feb 14 12:03:19 crc kubenswrapper[4904]: I0214 12:03:19.314079 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c3d261f-9b09-48b3-b47a-fcbc4985b47f-utilities\") pod \"certified-operators-q6rh2\" (UID: \"2c3d261f-9b09-48b3-b47a-fcbc4985b47f\") " pod="openshift-marketplace/certified-operators-q6rh2" Feb 14 12:03:19 crc kubenswrapper[4904]: I0214 12:03:19.314806 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c3d261f-9b09-48b3-b47a-fcbc4985b47f-catalog-content\") pod \"certified-operators-q6rh2\" (UID: \"2c3d261f-9b09-48b3-b47a-fcbc4985b47f\") " pod="openshift-marketplace/certified-operators-q6rh2" Feb 14 12:03:19 crc kubenswrapper[4904]: I0214 12:03:19.314896 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c3d261f-9b09-48b3-b47a-fcbc4985b47f-utilities\") pod \"certified-operators-q6rh2\" (UID: \"2c3d261f-9b09-48b3-b47a-fcbc4985b47f\") " pod="openshift-marketplace/certified-operators-q6rh2" Feb 14 12:03:19 crc kubenswrapper[4904]: I0214 12:03:19.355245 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpj79\" (UniqueName: \"kubernetes.io/projected/2c3d261f-9b09-48b3-b47a-fcbc4985b47f-kube-api-access-wpj79\") pod \"certified-operators-q6rh2\" (UID: \"2c3d261f-9b09-48b3-b47a-fcbc4985b47f\") " pod="openshift-marketplace/certified-operators-q6rh2" Feb 14 12:03:19 crc kubenswrapper[4904]: I0214 12:03:19.433201 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6rh2" Feb 14 12:03:19 crc kubenswrapper[4904]: I0214 12:03:19.980243 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q6rh2"] Feb 14 12:03:20 crc kubenswrapper[4904]: I0214 12:03:20.672357 4904 generic.go:334] "Generic (PLEG): container finished" podID="2c3d261f-9b09-48b3-b47a-fcbc4985b47f" containerID="5a3ed5e72dfaa241b1056b1e42968bc2085e3b8259326c10c3b7282108e1ea41" exitCode=0 Feb 14 12:03:20 crc kubenswrapper[4904]: I0214 12:03:20.672418 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6rh2" event={"ID":"2c3d261f-9b09-48b3-b47a-fcbc4985b47f","Type":"ContainerDied","Data":"5a3ed5e72dfaa241b1056b1e42968bc2085e3b8259326c10c3b7282108e1ea41"} Feb 14 12:03:20 crc kubenswrapper[4904]: I0214 12:03:20.672737 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6rh2" event={"ID":"2c3d261f-9b09-48b3-b47a-fcbc4985b47f","Type":"ContainerStarted","Data":"48711adc0e9272cdbf1bffde5393e04c18a95fe585175c4f66531c1d6155a213"} Feb 14 12:03:22 crc kubenswrapper[4904]: I0214 12:03:22.690932 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6rh2" event={"ID":"2c3d261f-9b09-48b3-b47a-fcbc4985b47f","Type":"ContainerStarted","Data":"03a829310793edafeea7c8b091465b0fda85cfc3c42aea740aff292ff324a7e0"} Feb 14 12:03:23 crc kubenswrapper[4904]: I0214 12:03:23.701362 4904 generic.go:334] "Generic (PLEG): container finished" podID="2c3d261f-9b09-48b3-b47a-fcbc4985b47f" containerID="03a829310793edafeea7c8b091465b0fda85cfc3c42aea740aff292ff324a7e0" exitCode=0 Feb 14 12:03:23 crc kubenswrapper[4904]: I0214 12:03:23.701447 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6rh2" event={"ID":"2c3d261f-9b09-48b3-b47a-fcbc4985b47f","Type":"ContainerDied","Data":"03a829310793edafeea7c8b091465b0fda85cfc3c42aea740aff292ff324a7e0"} Feb 14 12:03:24 crc kubenswrapper[4904]: I0214 12:03:24.836074 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 12:03:24 crc kubenswrapper[4904]: E0214 12:03:24.836655 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:03:25 crc kubenswrapper[4904]: I0214 12:03:25.716878 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6rh2" event={"ID":"2c3d261f-9b09-48b3-b47a-fcbc4985b47f","Type":"ContainerStarted","Data":"5a00bf1e873e71b5595628cf46ab881e4b98daea4a9e23e77036f93c206ca907"} Feb 14 12:03:25 crc kubenswrapper[4904]: I0214 12:03:25.744002 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q6rh2" podStartSLOduration=2.690115179 podStartE2EDuration="6.743981823s" podCreationTimestamp="2026-02-14 12:03:19 +0000 UTC" firstStartedPulling="2026-02-14 12:03:20.675510904 +0000 UTC m=+3191.488275565" lastFinishedPulling="2026-02-14 12:03:24.729377548 +0000 UTC m=+3195.542142209" observedRunningTime="2026-02-14 12:03:25.74026008 +0000 UTC m=+3196.553024761" watchObservedRunningTime="2026-02-14 12:03:25.743981823 +0000 UTC m=+3196.556746484" Feb 14 12:03:29 crc kubenswrapper[4904]: I0214 12:03:29.435091 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q6rh2" Feb 14 12:03:29 crc kubenswrapper[4904]: I0214 12:03:29.435468 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q6rh2" Feb 14 12:03:29 crc kubenswrapper[4904]: I0214 12:03:29.494055 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q6rh2" Feb 14 12:03:29 crc kubenswrapper[4904]: I0214 12:03:29.808096 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q6rh2" Feb 14 12:03:30 crc kubenswrapper[4904]: I0214 12:03:30.486949 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q6rh2"] Feb 14 12:03:31 crc kubenswrapper[4904]: I0214 12:03:31.766738 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-q6rh2" podUID="2c3d261f-9b09-48b3-b47a-fcbc4985b47f" containerName="registry-server" containerID="cri-o://5a00bf1e873e71b5595628cf46ab881e4b98daea4a9e23e77036f93c206ca907" gracePeriod=2 Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.200478 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6rh2" Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.256117 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpj79\" (UniqueName: \"kubernetes.io/projected/2c3d261f-9b09-48b3-b47a-fcbc4985b47f-kube-api-access-wpj79\") pod \"2c3d261f-9b09-48b3-b47a-fcbc4985b47f\" (UID: \"2c3d261f-9b09-48b3-b47a-fcbc4985b47f\") " Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.256216 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c3d261f-9b09-48b3-b47a-fcbc4985b47f-utilities\") pod \"2c3d261f-9b09-48b3-b47a-fcbc4985b47f\" (UID: \"2c3d261f-9b09-48b3-b47a-fcbc4985b47f\") " Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.256373 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c3d261f-9b09-48b3-b47a-fcbc4985b47f-catalog-content\") pod \"2c3d261f-9b09-48b3-b47a-fcbc4985b47f\" (UID: \"2c3d261f-9b09-48b3-b47a-fcbc4985b47f\") " Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.256936 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c3d261f-9b09-48b3-b47a-fcbc4985b47f-utilities" (OuterVolumeSpecName: "utilities") pod "2c3d261f-9b09-48b3-b47a-fcbc4985b47f" (UID: "2c3d261f-9b09-48b3-b47a-fcbc4985b47f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.262459 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c3d261f-9b09-48b3-b47a-fcbc4985b47f-kube-api-access-wpj79" (OuterVolumeSpecName: "kube-api-access-wpj79") pod "2c3d261f-9b09-48b3-b47a-fcbc4985b47f" (UID: "2c3d261f-9b09-48b3-b47a-fcbc4985b47f"). InnerVolumeSpecName "kube-api-access-wpj79". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.305290 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c3d261f-9b09-48b3-b47a-fcbc4985b47f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c3d261f-9b09-48b3-b47a-fcbc4985b47f" (UID: "2c3d261f-9b09-48b3-b47a-fcbc4985b47f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.358451 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpj79\" (UniqueName: \"kubernetes.io/projected/2c3d261f-9b09-48b3-b47a-fcbc4985b47f-kube-api-access-wpj79\") on node \"crc\" DevicePath \"\"" Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.358486 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c3d261f-9b09-48b3-b47a-fcbc4985b47f-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.358496 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c3d261f-9b09-48b3-b47a-fcbc4985b47f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.775541 4904 generic.go:334] "Generic (PLEG): container finished" podID="2c3d261f-9b09-48b3-b47a-fcbc4985b47f" containerID="5a00bf1e873e71b5595628cf46ab881e4b98daea4a9e23e77036f93c206ca907" exitCode=0 Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.775586 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6rh2" event={"ID":"2c3d261f-9b09-48b3-b47a-fcbc4985b47f","Type":"ContainerDied","Data":"5a00bf1e873e71b5595628cf46ab881e4b98daea4a9e23e77036f93c206ca907"} Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.775678 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6rh2" event={"ID":"2c3d261f-9b09-48b3-b47a-fcbc4985b47f","Type":"ContainerDied","Data":"48711adc0e9272cdbf1bffde5393e04c18a95fe585175c4f66531c1d6155a213"} Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.775702 4904 scope.go:117] "RemoveContainer" containerID="5a00bf1e873e71b5595628cf46ab881e4b98daea4a9e23e77036f93c206ca907" Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.775848 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6rh2" Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.803039 4904 scope.go:117] "RemoveContainer" containerID="03a829310793edafeea7c8b091465b0fda85cfc3c42aea740aff292ff324a7e0" Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.821179 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q6rh2"] Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.828345 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-q6rh2"] Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.831445 4904 scope.go:117] "RemoveContainer" containerID="5a3ed5e72dfaa241b1056b1e42968bc2085e3b8259326c10c3b7282108e1ea41" Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.875238 4904 scope.go:117] "RemoveContainer" containerID="5a00bf1e873e71b5595628cf46ab881e4b98daea4a9e23e77036f93c206ca907" Feb 14 12:03:32 crc kubenswrapper[4904]: E0214 12:03:32.875734 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a00bf1e873e71b5595628cf46ab881e4b98daea4a9e23e77036f93c206ca907\": container with ID starting with 5a00bf1e873e71b5595628cf46ab881e4b98daea4a9e23e77036f93c206ca907 not found: ID does not exist" containerID="5a00bf1e873e71b5595628cf46ab881e4b98daea4a9e23e77036f93c206ca907" Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.875787 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a00bf1e873e71b5595628cf46ab881e4b98daea4a9e23e77036f93c206ca907"} err="failed to get container status \"5a00bf1e873e71b5595628cf46ab881e4b98daea4a9e23e77036f93c206ca907\": rpc error: code = NotFound desc = could not find container \"5a00bf1e873e71b5595628cf46ab881e4b98daea4a9e23e77036f93c206ca907\": container with ID starting with 5a00bf1e873e71b5595628cf46ab881e4b98daea4a9e23e77036f93c206ca907 not found: ID does not exist" Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.875822 4904 scope.go:117] "RemoveContainer" containerID="03a829310793edafeea7c8b091465b0fda85cfc3c42aea740aff292ff324a7e0" Feb 14 12:03:32 crc kubenswrapper[4904]: E0214 12:03:32.876731 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03a829310793edafeea7c8b091465b0fda85cfc3c42aea740aff292ff324a7e0\": container with ID starting with 03a829310793edafeea7c8b091465b0fda85cfc3c42aea740aff292ff324a7e0 not found: ID does not exist" containerID="03a829310793edafeea7c8b091465b0fda85cfc3c42aea740aff292ff324a7e0" Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.876798 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03a829310793edafeea7c8b091465b0fda85cfc3c42aea740aff292ff324a7e0"} err="failed to get container status \"03a829310793edafeea7c8b091465b0fda85cfc3c42aea740aff292ff324a7e0\": rpc error: code = NotFound desc = could not find container \"03a829310793edafeea7c8b091465b0fda85cfc3c42aea740aff292ff324a7e0\": container with ID starting with 03a829310793edafeea7c8b091465b0fda85cfc3c42aea740aff292ff324a7e0 not found: ID does not exist" Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.876876 4904 scope.go:117] "RemoveContainer" containerID="5a3ed5e72dfaa241b1056b1e42968bc2085e3b8259326c10c3b7282108e1ea41" Feb 14 12:03:32 crc kubenswrapper[4904]: E0214 12:03:32.877302 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a3ed5e72dfaa241b1056b1e42968bc2085e3b8259326c10c3b7282108e1ea41\": container with ID starting with 5a3ed5e72dfaa241b1056b1e42968bc2085e3b8259326c10c3b7282108e1ea41 not found: ID does not exist" containerID="5a3ed5e72dfaa241b1056b1e42968bc2085e3b8259326c10c3b7282108e1ea41" Feb 14 12:03:32 crc kubenswrapper[4904]: I0214 12:03:32.877346 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a3ed5e72dfaa241b1056b1e42968bc2085e3b8259326c10c3b7282108e1ea41"} err="failed to get container status \"5a3ed5e72dfaa241b1056b1e42968bc2085e3b8259326c10c3b7282108e1ea41\": rpc error: code = NotFound desc = could not find container \"5a3ed5e72dfaa241b1056b1e42968bc2085e3b8259326c10c3b7282108e1ea41\": container with ID starting with 5a3ed5e72dfaa241b1056b1e42968bc2085e3b8259326c10c3b7282108e1ea41 not found: ID does not exist" Feb 14 12:03:33 crc kubenswrapper[4904]: I0214 12:03:33.846391 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c3d261f-9b09-48b3-b47a-fcbc4985b47f" path="/var/lib/kubelet/pods/2c3d261f-9b09-48b3-b47a-fcbc4985b47f/volumes" Feb 14 12:03:37 crc kubenswrapper[4904]: I0214 12:03:37.836957 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 12:03:37 crc kubenswrapper[4904]: E0214 12:03:37.837761 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:03:52 crc kubenswrapper[4904]: I0214 12:03:52.836883 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 12:03:53 crc kubenswrapper[4904]: I0214 12:03:53.963014 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"5912324274aefd184db28263c4127b73102f7b54d640eb6b2392eb10f3b91ba3"} Feb 14 12:06:16 crc kubenswrapper[4904]: I0214 12:06:16.383729 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:06:16 crc kubenswrapper[4904]: I0214 12:06:16.384373 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:06:46 crc kubenswrapper[4904]: I0214 12:06:46.382890 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:06:46 crc kubenswrapper[4904]: I0214 12:06:46.383715 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:07:16 crc kubenswrapper[4904]: I0214 12:07:16.382416 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:07:16 crc kubenswrapper[4904]: I0214 12:07:16.382936 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:07:16 crc kubenswrapper[4904]: I0214 12:07:16.382978 4904 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 12:07:16 crc kubenswrapper[4904]: I0214 12:07:16.383603 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5912324274aefd184db28263c4127b73102f7b54d640eb6b2392eb10f3b91ba3"} pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 14 12:07:16 crc kubenswrapper[4904]: I0214 12:07:16.383660 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" containerID="cri-o://5912324274aefd184db28263c4127b73102f7b54d640eb6b2392eb10f3b91ba3" gracePeriod=600 Feb 14 12:07:16 crc kubenswrapper[4904]: I0214 12:07:16.698744 4904 generic.go:334] "Generic (PLEG): container finished" podID="7790d276-2506-4641-8c22-67e6441417f0" containerID="5912324274aefd184db28263c4127b73102f7b54d640eb6b2392eb10f3b91ba3" exitCode=0 Feb 14 12:07:16 crc kubenswrapper[4904]: I0214 12:07:16.698819 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerDied","Data":"5912324274aefd184db28263c4127b73102f7b54d640eb6b2392eb10f3b91ba3"} Feb 14 12:07:16 crc kubenswrapper[4904]: I0214 12:07:16.699399 4904 scope.go:117] "RemoveContainer" containerID="0f5ee3e9ff8d74ca4e696e0ee786878aa56a08c2eb5b8483fc37c2af61118067" Feb 14 12:07:17 crc kubenswrapper[4904]: I0214 12:07:17.714449 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664"} Feb 14 12:08:03 crc kubenswrapper[4904]: I0214 12:08:03.077483 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nlvxx"] Feb 14 12:08:03 crc kubenswrapper[4904]: E0214 12:08:03.078479 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c3d261f-9b09-48b3-b47a-fcbc4985b47f" containerName="registry-server" Feb 14 12:08:03 crc kubenswrapper[4904]: I0214 12:08:03.078493 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c3d261f-9b09-48b3-b47a-fcbc4985b47f" containerName="registry-server" Feb 14 12:08:03 crc kubenswrapper[4904]: E0214 12:08:03.078536 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c3d261f-9b09-48b3-b47a-fcbc4985b47f" containerName="extract-utilities" Feb 14 12:08:03 crc kubenswrapper[4904]: I0214 12:08:03.078545 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c3d261f-9b09-48b3-b47a-fcbc4985b47f" containerName="extract-utilities" Feb 14 12:08:03 crc kubenswrapper[4904]: E0214 12:08:03.078562 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c3d261f-9b09-48b3-b47a-fcbc4985b47f" containerName="extract-content" Feb 14 12:08:03 crc kubenswrapper[4904]: I0214 12:08:03.078571 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c3d261f-9b09-48b3-b47a-fcbc4985b47f" containerName="extract-content" Feb 14 12:08:03 crc kubenswrapper[4904]: I0214 12:08:03.078810 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c3d261f-9b09-48b3-b47a-fcbc4985b47f" containerName="registry-server" Feb 14 12:08:03 crc kubenswrapper[4904]: I0214 12:08:03.080536 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nlvxx" Feb 14 12:08:03 crc kubenswrapper[4904]: I0214 12:08:03.087038 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nlvxx"] Feb 14 12:08:03 crc kubenswrapper[4904]: I0214 12:08:03.170224 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng8bn\" (UniqueName: \"kubernetes.io/projected/9d9abccf-ebf8-4692-b7ed-d416c8d8760a-kube-api-access-ng8bn\") pod \"redhat-operators-nlvxx\" (UID: \"9d9abccf-ebf8-4692-b7ed-d416c8d8760a\") " pod="openshift-marketplace/redhat-operators-nlvxx" Feb 14 12:08:03 crc kubenswrapper[4904]: I0214 12:08:03.170351 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d9abccf-ebf8-4692-b7ed-d416c8d8760a-utilities\") pod \"redhat-operators-nlvxx\" (UID: \"9d9abccf-ebf8-4692-b7ed-d416c8d8760a\") " pod="openshift-marketplace/redhat-operators-nlvxx" Feb 14 12:08:03 crc kubenswrapper[4904]: I0214 12:08:03.170382 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d9abccf-ebf8-4692-b7ed-d416c8d8760a-catalog-content\") pod \"redhat-operators-nlvxx\" (UID: \"9d9abccf-ebf8-4692-b7ed-d416c8d8760a\") " pod="openshift-marketplace/redhat-operators-nlvxx" Feb 14 12:08:03 crc kubenswrapper[4904]: I0214 12:08:03.271884 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d9abccf-ebf8-4692-b7ed-d416c8d8760a-utilities\") pod \"redhat-operators-nlvxx\" (UID: \"9d9abccf-ebf8-4692-b7ed-d416c8d8760a\") " pod="openshift-marketplace/redhat-operators-nlvxx" Feb 14 12:08:03 crc kubenswrapper[4904]: I0214 12:08:03.271954 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d9abccf-ebf8-4692-b7ed-d416c8d8760a-catalog-content\") pod \"redhat-operators-nlvxx\" (UID: \"9d9abccf-ebf8-4692-b7ed-d416c8d8760a\") " pod="openshift-marketplace/redhat-operators-nlvxx" Feb 14 12:08:03 crc kubenswrapper[4904]: I0214 12:08:03.272107 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng8bn\" (UniqueName: \"kubernetes.io/projected/9d9abccf-ebf8-4692-b7ed-d416c8d8760a-kube-api-access-ng8bn\") pod \"redhat-operators-nlvxx\" (UID: \"9d9abccf-ebf8-4692-b7ed-d416c8d8760a\") " pod="openshift-marketplace/redhat-operators-nlvxx" Feb 14 12:08:03 crc kubenswrapper[4904]: I0214 12:08:03.273000 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d9abccf-ebf8-4692-b7ed-d416c8d8760a-utilities\") pod \"redhat-operators-nlvxx\" (UID: \"9d9abccf-ebf8-4692-b7ed-d416c8d8760a\") " pod="openshift-marketplace/redhat-operators-nlvxx" Feb 14 12:08:03 crc kubenswrapper[4904]: I0214 12:08:03.273180 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d9abccf-ebf8-4692-b7ed-d416c8d8760a-catalog-content\") pod \"redhat-operators-nlvxx\" (UID: \"9d9abccf-ebf8-4692-b7ed-d416c8d8760a\") " pod="openshift-marketplace/redhat-operators-nlvxx" Feb 14 12:08:03 crc kubenswrapper[4904]: I0214 12:08:03.290333 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng8bn\" (UniqueName: \"kubernetes.io/projected/9d9abccf-ebf8-4692-b7ed-d416c8d8760a-kube-api-access-ng8bn\") pod \"redhat-operators-nlvxx\" (UID: \"9d9abccf-ebf8-4692-b7ed-d416c8d8760a\") " pod="openshift-marketplace/redhat-operators-nlvxx" Feb 14 12:08:03 crc kubenswrapper[4904]: I0214 12:08:03.407533 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nlvxx" Feb 14 12:08:03 crc kubenswrapper[4904]: I0214 12:08:03.924464 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nlvxx"] Feb 14 12:08:04 crc kubenswrapper[4904]: I0214 12:08:04.099096 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nlvxx" event={"ID":"9d9abccf-ebf8-4692-b7ed-d416c8d8760a","Type":"ContainerStarted","Data":"90d9f5c96c4789db3e6c22e939c7a0de6f348d963962b5743e80dc848d07fe15"} Feb 14 12:08:05 crc kubenswrapper[4904]: I0214 12:08:05.112487 4904 generic.go:334] "Generic (PLEG): container finished" podID="9d9abccf-ebf8-4692-b7ed-d416c8d8760a" containerID="7ab66e4b89928ac3a7d90669636498a7f7cadd4f2100df4bb1441f2499aa41f0" exitCode=0 Feb 14 12:08:05 crc kubenswrapper[4904]: I0214 12:08:05.112667 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nlvxx" event={"ID":"9d9abccf-ebf8-4692-b7ed-d416c8d8760a","Type":"ContainerDied","Data":"7ab66e4b89928ac3a7d90669636498a7f7cadd4f2100df4bb1441f2499aa41f0"} Feb 14 12:08:05 crc kubenswrapper[4904]: I0214 12:08:05.115247 4904 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 14 12:08:06 crc kubenswrapper[4904]: I0214 12:08:06.124154 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nlvxx" event={"ID":"9d9abccf-ebf8-4692-b7ed-d416c8d8760a","Type":"ContainerStarted","Data":"6188d17a681609f1c421e7a1beefa934661d2b1a8b4b151c4ad182c4cf2e9be8"} Feb 14 12:08:11 crc kubenswrapper[4904]: I0214 12:08:11.175007 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nlvxx" event={"ID":"9d9abccf-ebf8-4692-b7ed-d416c8d8760a","Type":"ContainerDied","Data":"6188d17a681609f1c421e7a1beefa934661d2b1a8b4b151c4ad182c4cf2e9be8"} Feb 14 12:08:11 crc kubenswrapper[4904]: I0214 12:08:11.175095 4904 generic.go:334] "Generic (PLEG): container finished" podID="9d9abccf-ebf8-4692-b7ed-d416c8d8760a" containerID="6188d17a681609f1c421e7a1beefa934661d2b1a8b4b151c4ad182c4cf2e9be8" exitCode=0 Feb 14 12:08:12 crc kubenswrapper[4904]: I0214 12:08:12.188962 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nlvxx" event={"ID":"9d9abccf-ebf8-4692-b7ed-d416c8d8760a","Type":"ContainerStarted","Data":"04daf5f8596687ff4c9a8cdc070deb8bd92cf7b1dd34385947c7f0166b19e939"} Feb 14 12:08:12 crc kubenswrapper[4904]: I0214 12:08:12.216823 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nlvxx" podStartSLOduration=2.647914545 podStartE2EDuration="9.216800892s" podCreationTimestamp="2026-02-14 12:08:03 +0000 UTC" firstStartedPulling="2026-02-14 12:08:05.115026191 +0000 UTC m=+3475.927790852" lastFinishedPulling="2026-02-14 12:08:11.683912518 +0000 UTC m=+3482.496677199" observedRunningTime="2026-02-14 12:08:12.209020449 +0000 UTC m=+3483.021785100" watchObservedRunningTime="2026-02-14 12:08:12.216800892 +0000 UTC m=+3483.029565553" Feb 14 12:08:13 crc kubenswrapper[4904]: I0214 12:08:13.407633 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nlvxx" Feb 14 12:08:13 crc kubenswrapper[4904]: I0214 12:08:13.408010 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nlvxx" Feb 14 12:08:14 crc kubenswrapper[4904]: I0214 12:08:14.457239 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nlvxx" podUID="9d9abccf-ebf8-4692-b7ed-d416c8d8760a" containerName="registry-server" probeResult="failure" output=< Feb 14 12:08:14 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 12:08:14 crc kubenswrapper[4904]: > Feb 14 12:08:24 crc kubenswrapper[4904]: I0214 12:08:24.450281 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nlvxx" podUID="9d9abccf-ebf8-4692-b7ed-d416c8d8760a" containerName="registry-server" probeResult="failure" output=< Feb 14 12:08:24 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 12:08:24 crc kubenswrapper[4904]: > Feb 14 12:08:34 crc kubenswrapper[4904]: I0214 12:08:34.450685 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nlvxx" podUID="9d9abccf-ebf8-4692-b7ed-d416c8d8760a" containerName="registry-server" probeResult="failure" output=< Feb 14 12:08:34 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 12:08:34 crc kubenswrapper[4904]: > Feb 14 12:08:43 crc kubenswrapper[4904]: I0214 12:08:43.460437 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nlvxx" Feb 14 12:08:43 crc kubenswrapper[4904]: I0214 12:08:43.507514 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nlvxx" Feb 14 12:08:43 crc kubenswrapper[4904]: I0214 12:08:43.694646 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nlvxx"] Feb 14 12:08:45 crc kubenswrapper[4904]: I0214 12:08:45.487611 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nlvxx" podUID="9d9abccf-ebf8-4692-b7ed-d416c8d8760a" containerName="registry-server" containerID="cri-o://04daf5f8596687ff4c9a8cdc070deb8bd92cf7b1dd34385947c7f0166b19e939" gracePeriod=2 Feb 14 12:08:45 crc kubenswrapper[4904]: I0214 12:08:45.954324 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nlvxx" Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.065524 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d9abccf-ebf8-4692-b7ed-d416c8d8760a-catalog-content\") pod \"9d9abccf-ebf8-4692-b7ed-d416c8d8760a\" (UID: \"9d9abccf-ebf8-4692-b7ed-d416c8d8760a\") " Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.065597 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ng8bn\" (UniqueName: \"kubernetes.io/projected/9d9abccf-ebf8-4692-b7ed-d416c8d8760a-kube-api-access-ng8bn\") pod \"9d9abccf-ebf8-4692-b7ed-d416c8d8760a\" (UID: \"9d9abccf-ebf8-4692-b7ed-d416c8d8760a\") " Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.066771 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d9abccf-ebf8-4692-b7ed-d416c8d8760a-utilities\") pod \"9d9abccf-ebf8-4692-b7ed-d416c8d8760a\" (UID: \"9d9abccf-ebf8-4692-b7ed-d416c8d8760a\") " Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.067410 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d9abccf-ebf8-4692-b7ed-d416c8d8760a-utilities" (OuterVolumeSpecName: "utilities") pod "9d9abccf-ebf8-4692-b7ed-d416c8d8760a" (UID: "9d9abccf-ebf8-4692-b7ed-d416c8d8760a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.078814 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d9abccf-ebf8-4692-b7ed-d416c8d8760a-kube-api-access-ng8bn" (OuterVolumeSpecName: "kube-api-access-ng8bn") pod "9d9abccf-ebf8-4692-b7ed-d416c8d8760a" (UID: "9d9abccf-ebf8-4692-b7ed-d416c8d8760a"). InnerVolumeSpecName "kube-api-access-ng8bn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.169158 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ng8bn\" (UniqueName: \"kubernetes.io/projected/9d9abccf-ebf8-4692-b7ed-d416c8d8760a-kube-api-access-ng8bn\") on node \"crc\" DevicePath \"\"" Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.169201 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d9abccf-ebf8-4692-b7ed-d416c8d8760a-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.192519 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d9abccf-ebf8-4692-b7ed-d416c8d8760a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9d9abccf-ebf8-4692-b7ed-d416c8d8760a" (UID: "9d9abccf-ebf8-4692-b7ed-d416c8d8760a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.270929 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d9abccf-ebf8-4692-b7ed-d416c8d8760a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.504981 4904 generic.go:334] "Generic (PLEG): container finished" podID="9d9abccf-ebf8-4692-b7ed-d416c8d8760a" containerID="04daf5f8596687ff4c9a8cdc070deb8bd92cf7b1dd34385947c7f0166b19e939" exitCode=0 Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.505022 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nlvxx" event={"ID":"9d9abccf-ebf8-4692-b7ed-d416c8d8760a","Type":"ContainerDied","Data":"04daf5f8596687ff4c9a8cdc070deb8bd92cf7b1dd34385947c7f0166b19e939"} Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.505044 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nlvxx" Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.505061 4904 scope.go:117] "RemoveContainer" containerID="04daf5f8596687ff4c9a8cdc070deb8bd92cf7b1dd34385947c7f0166b19e939" Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.505049 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nlvxx" event={"ID":"9d9abccf-ebf8-4692-b7ed-d416c8d8760a","Type":"ContainerDied","Data":"90d9f5c96c4789db3e6c22e939c7a0de6f348d963962b5743e80dc848d07fe15"} Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.546791 4904 scope.go:117] "RemoveContainer" containerID="6188d17a681609f1c421e7a1beefa934661d2b1a8b4b151c4ad182c4cf2e9be8" Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.548080 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nlvxx"] Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.557334 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nlvxx"] Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.577045 4904 scope.go:117] "RemoveContainer" containerID="7ab66e4b89928ac3a7d90669636498a7f7cadd4f2100df4bb1441f2499aa41f0" Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.620537 4904 scope.go:117] "RemoveContainer" containerID="04daf5f8596687ff4c9a8cdc070deb8bd92cf7b1dd34385947c7f0166b19e939" Feb 14 12:08:46 crc kubenswrapper[4904]: E0214 12:08:46.621062 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04daf5f8596687ff4c9a8cdc070deb8bd92cf7b1dd34385947c7f0166b19e939\": container with ID starting with 04daf5f8596687ff4c9a8cdc070deb8bd92cf7b1dd34385947c7f0166b19e939 not found: ID does not exist" containerID="04daf5f8596687ff4c9a8cdc070deb8bd92cf7b1dd34385947c7f0166b19e939" Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.621116 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04daf5f8596687ff4c9a8cdc070deb8bd92cf7b1dd34385947c7f0166b19e939"} err="failed to get container status \"04daf5f8596687ff4c9a8cdc070deb8bd92cf7b1dd34385947c7f0166b19e939\": rpc error: code = NotFound desc = could not find container \"04daf5f8596687ff4c9a8cdc070deb8bd92cf7b1dd34385947c7f0166b19e939\": container with ID starting with 04daf5f8596687ff4c9a8cdc070deb8bd92cf7b1dd34385947c7f0166b19e939 not found: ID does not exist" Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.621148 4904 scope.go:117] "RemoveContainer" containerID="6188d17a681609f1c421e7a1beefa934661d2b1a8b4b151c4ad182c4cf2e9be8" Feb 14 12:08:46 crc kubenswrapper[4904]: E0214 12:08:46.621598 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6188d17a681609f1c421e7a1beefa934661d2b1a8b4b151c4ad182c4cf2e9be8\": container with ID starting with 6188d17a681609f1c421e7a1beefa934661d2b1a8b4b151c4ad182c4cf2e9be8 not found: ID does not exist" containerID="6188d17a681609f1c421e7a1beefa934661d2b1a8b4b151c4ad182c4cf2e9be8" Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.621651 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6188d17a681609f1c421e7a1beefa934661d2b1a8b4b151c4ad182c4cf2e9be8"} err="failed to get container status \"6188d17a681609f1c421e7a1beefa934661d2b1a8b4b151c4ad182c4cf2e9be8\": rpc error: code = NotFound desc = could not find container \"6188d17a681609f1c421e7a1beefa934661d2b1a8b4b151c4ad182c4cf2e9be8\": container with ID starting with 6188d17a681609f1c421e7a1beefa934661d2b1a8b4b151c4ad182c4cf2e9be8 not found: ID does not exist" Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.621683 4904 scope.go:117] "RemoveContainer" containerID="7ab66e4b89928ac3a7d90669636498a7f7cadd4f2100df4bb1441f2499aa41f0" Feb 14 12:08:46 crc kubenswrapper[4904]: E0214 12:08:46.622107 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ab66e4b89928ac3a7d90669636498a7f7cadd4f2100df4bb1441f2499aa41f0\": container with ID starting with 7ab66e4b89928ac3a7d90669636498a7f7cadd4f2100df4bb1441f2499aa41f0 not found: ID does not exist" containerID="7ab66e4b89928ac3a7d90669636498a7f7cadd4f2100df4bb1441f2499aa41f0" Feb 14 12:08:46 crc kubenswrapper[4904]: I0214 12:08:46.622137 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ab66e4b89928ac3a7d90669636498a7f7cadd4f2100df4bb1441f2499aa41f0"} err="failed to get container status \"7ab66e4b89928ac3a7d90669636498a7f7cadd4f2100df4bb1441f2499aa41f0\": rpc error: code = NotFound desc = could not find container \"7ab66e4b89928ac3a7d90669636498a7f7cadd4f2100df4bb1441f2499aa41f0\": container with ID starting with 7ab66e4b89928ac3a7d90669636498a7f7cadd4f2100df4bb1441f2499aa41f0 not found: ID does not exist" Feb 14 12:08:47 crc kubenswrapper[4904]: I0214 12:08:47.850394 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d9abccf-ebf8-4692-b7ed-d416c8d8760a" path="/var/lib/kubelet/pods/9d9abccf-ebf8-4692-b7ed-d416c8d8760a/volumes" Feb 14 12:09:16 crc kubenswrapper[4904]: I0214 12:09:16.383336 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:09:16 crc kubenswrapper[4904]: I0214 12:09:16.383874 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:09:46 crc kubenswrapper[4904]: I0214 12:09:46.382870 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:09:46 crc kubenswrapper[4904]: I0214 12:09:46.383539 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:10:16 crc kubenswrapper[4904]: I0214 12:10:16.383688 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:10:16 crc kubenswrapper[4904]: I0214 12:10:16.384426 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:10:16 crc kubenswrapper[4904]: I0214 12:10:16.384499 4904 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 12:10:16 crc kubenswrapper[4904]: I0214 12:10:16.385756 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664"} pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 14 12:10:16 crc kubenswrapper[4904]: I0214 12:10:16.385887 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" containerID="cri-o://38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" gracePeriod=600 Feb 14 12:10:16 crc kubenswrapper[4904]: E0214 12:10:16.518166 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:10:17 crc kubenswrapper[4904]: I0214 12:10:17.229316 4904 generic.go:334] "Generic (PLEG): container finished" podID="7790d276-2506-4641-8c22-67e6441417f0" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" exitCode=0 Feb 14 12:10:17 crc kubenswrapper[4904]: I0214 12:10:17.229399 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerDied","Data":"38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664"} Feb 14 12:10:17 crc kubenswrapper[4904]: I0214 12:10:17.229630 4904 scope.go:117] "RemoveContainer" containerID="5912324274aefd184db28263c4127b73102f7b54d640eb6b2392eb10f3b91ba3" Feb 14 12:10:17 crc kubenswrapper[4904]: I0214 12:10:17.230420 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:10:17 crc kubenswrapper[4904]: E0214 12:10:17.230724 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:10:29 crc kubenswrapper[4904]: I0214 12:10:29.849381 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:10:29 crc kubenswrapper[4904]: E0214 12:10:29.850364 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:10:40 crc kubenswrapper[4904]: I0214 12:10:40.837121 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:10:40 crc kubenswrapper[4904]: E0214 12:10:40.839137 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:10:52 crc kubenswrapper[4904]: I0214 12:10:52.836128 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:10:52 crc kubenswrapper[4904]: E0214 12:10:52.836927 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:11:03 crc kubenswrapper[4904]: I0214 12:11:03.837702 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:11:03 crc kubenswrapper[4904]: E0214 12:11:03.839195 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:11:17 crc kubenswrapper[4904]: I0214 12:11:17.836737 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:11:17 crc kubenswrapper[4904]: E0214 12:11:17.838131 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:11:32 crc kubenswrapper[4904]: I0214 12:11:32.836491 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:11:32 crc kubenswrapper[4904]: E0214 12:11:32.837164 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:11:47 crc kubenswrapper[4904]: I0214 12:11:47.836786 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:11:47 crc kubenswrapper[4904]: E0214 12:11:47.837429 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:12:00 crc kubenswrapper[4904]: I0214 12:12:00.836600 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:12:00 crc kubenswrapper[4904]: E0214 12:12:00.837306 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:12:15 crc kubenswrapper[4904]: I0214 12:12:15.836424 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:12:15 crc kubenswrapper[4904]: E0214 12:12:15.837468 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:12:27 crc kubenswrapper[4904]: I0214 12:12:27.837789 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:12:27 crc kubenswrapper[4904]: E0214 12:12:27.838802 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:12:34 crc kubenswrapper[4904]: I0214 12:12:34.692863 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h68jl"] Feb 14 12:12:34 crc kubenswrapper[4904]: E0214 12:12:34.693472 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d9abccf-ebf8-4692-b7ed-d416c8d8760a" containerName="extract-utilities" Feb 14 12:12:34 crc kubenswrapper[4904]: I0214 12:12:34.693487 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d9abccf-ebf8-4692-b7ed-d416c8d8760a" containerName="extract-utilities" Feb 14 12:12:34 crc kubenswrapper[4904]: E0214 12:12:34.693505 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d9abccf-ebf8-4692-b7ed-d416c8d8760a" containerName="extract-content" Feb 14 12:12:34 crc kubenswrapper[4904]: I0214 12:12:34.693513 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d9abccf-ebf8-4692-b7ed-d416c8d8760a" containerName="extract-content" Feb 14 12:12:34 crc kubenswrapper[4904]: E0214 12:12:34.693533 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d9abccf-ebf8-4692-b7ed-d416c8d8760a" containerName="registry-server" Feb 14 12:12:34 crc kubenswrapper[4904]: I0214 12:12:34.693541 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d9abccf-ebf8-4692-b7ed-d416c8d8760a" containerName="registry-server" Feb 14 12:12:34 crc kubenswrapper[4904]: I0214 12:12:34.693766 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d9abccf-ebf8-4692-b7ed-d416c8d8760a" containerName="registry-server" Feb 14 12:12:34 crc kubenswrapper[4904]: I0214 12:12:34.695282 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h68jl" Feb 14 12:12:34 crc kubenswrapper[4904]: I0214 12:12:34.709029 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h68jl"] Feb 14 12:12:34 crc kubenswrapper[4904]: I0214 12:12:34.730086 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db7bd378-cb94-4ae1-b252-420f751b87bf-catalog-content\") pod \"redhat-marketplace-h68jl\" (UID: \"db7bd378-cb94-4ae1-b252-420f751b87bf\") " pod="openshift-marketplace/redhat-marketplace-h68jl" Feb 14 12:12:34 crc kubenswrapper[4904]: I0214 12:12:34.730153 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db7bd378-cb94-4ae1-b252-420f751b87bf-utilities\") pod \"redhat-marketplace-h68jl\" (UID: \"db7bd378-cb94-4ae1-b252-420f751b87bf\") " pod="openshift-marketplace/redhat-marketplace-h68jl" Feb 14 12:12:34 crc kubenswrapper[4904]: I0214 12:12:34.730188 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xj76\" (UniqueName: \"kubernetes.io/projected/db7bd378-cb94-4ae1-b252-420f751b87bf-kube-api-access-6xj76\") pod \"redhat-marketplace-h68jl\" (UID: \"db7bd378-cb94-4ae1-b252-420f751b87bf\") " pod="openshift-marketplace/redhat-marketplace-h68jl" Feb 14 12:12:34 crc kubenswrapper[4904]: I0214 12:12:34.831625 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db7bd378-cb94-4ae1-b252-420f751b87bf-catalog-content\") pod \"redhat-marketplace-h68jl\" (UID: \"db7bd378-cb94-4ae1-b252-420f751b87bf\") " pod="openshift-marketplace/redhat-marketplace-h68jl" Feb 14 12:12:34 crc kubenswrapper[4904]: I0214 12:12:34.831980 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db7bd378-cb94-4ae1-b252-420f751b87bf-utilities\") pod \"redhat-marketplace-h68jl\" (UID: \"db7bd378-cb94-4ae1-b252-420f751b87bf\") " pod="openshift-marketplace/redhat-marketplace-h68jl" Feb 14 12:12:34 crc kubenswrapper[4904]: I0214 12:12:34.832092 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xj76\" (UniqueName: \"kubernetes.io/projected/db7bd378-cb94-4ae1-b252-420f751b87bf-kube-api-access-6xj76\") pod \"redhat-marketplace-h68jl\" (UID: \"db7bd378-cb94-4ae1-b252-420f751b87bf\") " pod="openshift-marketplace/redhat-marketplace-h68jl" Feb 14 12:12:34 crc kubenswrapper[4904]: I0214 12:12:34.832447 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db7bd378-cb94-4ae1-b252-420f751b87bf-utilities\") pod \"redhat-marketplace-h68jl\" (UID: \"db7bd378-cb94-4ae1-b252-420f751b87bf\") " pod="openshift-marketplace/redhat-marketplace-h68jl" Feb 14 12:12:34 crc kubenswrapper[4904]: I0214 12:12:34.832709 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db7bd378-cb94-4ae1-b252-420f751b87bf-catalog-content\") pod \"redhat-marketplace-h68jl\" (UID: \"db7bd378-cb94-4ae1-b252-420f751b87bf\") " pod="openshift-marketplace/redhat-marketplace-h68jl" Feb 14 12:12:34 crc kubenswrapper[4904]: I0214 12:12:34.854023 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xj76\" (UniqueName: \"kubernetes.io/projected/db7bd378-cb94-4ae1-b252-420f751b87bf-kube-api-access-6xj76\") pod \"redhat-marketplace-h68jl\" (UID: \"db7bd378-cb94-4ae1-b252-420f751b87bf\") " pod="openshift-marketplace/redhat-marketplace-h68jl" Feb 14 12:12:35 crc kubenswrapper[4904]: I0214 12:12:35.013788 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h68jl" Feb 14 12:12:35 crc kubenswrapper[4904]: I0214 12:12:35.469925 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h68jl"] Feb 14 12:12:36 crc kubenswrapper[4904]: I0214 12:12:36.365680 4904 generic.go:334] "Generic (PLEG): container finished" podID="db7bd378-cb94-4ae1-b252-420f751b87bf" containerID="b47818f45fe7992f792e971d5c9ca786d1571c1207c7d7012fd2e4aa258b2853" exitCode=0 Feb 14 12:12:36 crc kubenswrapper[4904]: I0214 12:12:36.367315 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h68jl" event={"ID":"db7bd378-cb94-4ae1-b252-420f751b87bf","Type":"ContainerDied","Data":"b47818f45fe7992f792e971d5c9ca786d1571c1207c7d7012fd2e4aa258b2853"} Feb 14 12:12:36 crc kubenswrapper[4904]: I0214 12:12:36.367461 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h68jl" event={"ID":"db7bd378-cb94-4ae1-b252-420f751b87bf","Type":"ContainerStarted","Data":"52f1fa345169d2b5856761cbd126b640c61c3dc7fe49af22bb13c377a489be8c"} Feb 14 12:12:37 crc kubenswrapper[4904]: I0214 12:12:37.377080 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h68jl" event={"ID":"db7bd378-cb94-4ae1-b252-420f751b87bf","Type":"ContainerStarted","Data":"e57122af2d923f7eea52c61db535b0cca385e0f6435f35136aca86e376a5c09a"} Feb 14 12:12:37 crc kubenswrapper[4904]: E0214 12:12:37.950402 4904 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb7bd378_cb94_4ae1_b252_420f751b87bf.slice/crio-conmon-e57122af2d923f7eea52c61db535b0cca385e0f6435f35136aca86e376a5c09a.scope\": RecentStats: unable to find data in memory cache]" Feb 14 12:12:38 crc kubenswrapper[4904]: I0214 12:12:38.389631 4904 generic.go:334] "Generic (PLEG): container finished" podID="db7bd378-cb94-4ae1-b252-420f751b87bf" containerID="e57122af2d923f7eea52c61db535b0cca385e0f6435f35136aca86e376a5c09a" exitCode=0 Feb 14 12:12:38 crc kubenswrapper[4904]: I0214 12:12:38.389705 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h68jl" event={"ID":"db7bd378-cb94-4ae1-b252-420f751b87bf","Type":"ContainerDied","Data":"e57122af2d923f7eea52c61db535b0cca385e0f6435f35136aca86e376a5c09a"} Feb 14 12:12:38 crc kubenswrapper[4904]: I0214 12:12:38.836402 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:12:38 crc kubenswrapper[4904]: E0214 12:12:38.836688 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:12:39 crc kubenswrapper[4904]: I0214 12:12:39.398657 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h68jl" event={"ID":"db7bd378-cb94-4ae1-b252-420f751b87bf","Type":"ContainerStarted","Data":"7defe3fe51e1cf297682fe42bae8b508e126ad9cd668347760f25f76796dbccd"} Feb 14 12:12:45 crc kubenswrapper[4904]: I0214 12:12:45.015078 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h68jl" Feb 14 12:12:45 crc kubenswrapper[4904]: I0214 12:12:45.017778 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h68jl" Feb 14 12:12:45 crc kubenswrapper[4904]: I0214 12:12:45.066603 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h68jl" Feb 14 12:12:45 crc kubenswrapper[4904]: I0214 12:12:45.088659 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h68jl" podStartSLOduration=8.657432725 podStartE2EDuration="11.088636124s" podCreationTimestamp="2026-02-14 12:12:34 +0000 UTC" firstStartedPulling="2026-02-14 12:12:36.369413869 +0000 UTC m=+3747.182178530" lastFinishedPulling="2026-02-14 12:12:38.800617258 +0000 UTC m=+3749.613381929" observedRunningTime="2026-02-14 12:12:39.418543095 +0000 UTC m=+3750.231307766" watchObservedRunningTime="2026-02-14 12:12:45.088636124 +0000 UTC m=+3755.901400795" Feb 14 12:12:45 crc kubenswrapper[4904]: I0214 12:12:45.487275 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h68jl" Feb 14 12:12:45 crc kubenswrapper[4904]: I0214 12:12:45.542695 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h68jl"] Feb 14 12:12:47 crc kubenswrapper[4904]: I0214 12:12:47.454792 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h68jl" podUID="db7bd378-cb94-4ae1-b252-420f751b87bf" containerName="registry-server" containerID="cri-o://7defe3fe51e1cf297682fe42bae8b508e126ad9cd668347760f25f76796dbccd" gracePeriod=2 Feb 14 12:12:47 crc kubenswrapper[4904]: I0214 12:12:47.871546 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h68jl" Feb 14 12:12:47 crc kubenswrapper[4904]: I0214 12:12:47.973970 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db7bd378-cb94-4ae1-b252-420f751b87bf-catalog-content\") pod \"db7bd378-cb94-4ae1-b252-420f751b87bf\" (UID: \"db7bd378-cb94-4ae1-b252-420f751b87bf\") " Feb 14 12:12:47 crc kubenswrapper[4904]: I0214 12:12:47.974019 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xj76\" (UniqueName: \"kubernetes.io/projected/db7bd378-cb94-4ae1-b252-420f751b87bf-kube-api-access-6xj76\") pod \"db7bd378-cb94-4ae1-b252-420f751b87bf\" (UID: \"db7bd378-cb94-4ae1-b252-420f751b87bf\") " Feb 14 12:12:47 crc kubenswrapper[4904]: I0214 12:12:47.974081 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db7bd378-cb94-4ae1-b252-420f751b87bf-utilities\") pod \"db7bd378-cb94-4ae1-b252-420f751b87bf\" (UID: \"db7bd378-cb94-4ae1-b252-420f751b87bf\") " Feb 14 12:12:47 crc kubenswrapper[4904]: I0214 12:12:47.977129 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db7bd378-cb94-4ae1-b252-420f751b87bf-utilities" (OuterVolumeSpecName: "utilities") pod "db7bd378-cb94-4ae1-b252-420f751b87bf" (UID: "db7bd378-cb94-4ae1-b252-420f751b87bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:12:47 crc kubenswrapper[4904]: I0214 12:12:47.994872 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db7bd378-cb94-4ae1-b252-420f751b87bf-kube-api-access-6xj76" (OuterVolumeSpecName: "kube-api-access-6xj76") pod "db7bd378-cb94-4ae1-b252-420f751b87bf" (UID: "db7bd378-cb94-4ae1-b252-420f751b87bf"). InnerVolumeSpecName "kube-api-access-6xj76". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:12:48 crc kubenswrapper[4904]: I0214 12:12:48.015105 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db7bd378-cb94-4ae1-b252-420f751b87bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db7bd378-cb94-4ae1-b252-420f751b87bf" (UID: "db7bd378-cb94-4ae1-b252-420f751b87bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:12:48 crc kubenswrapper[4904]: I0214 12:12:48.075636 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db7bd378-cb94-4ae1-b252-420f751b87bf-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 12:12:48 crc kubenswrapper[4904]: I0214 12:12:48.075666 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xj76\" (UniqueName: \"kubernetes.io/projected/db7bd378-cb94-4ae1-b252-420f751b87bf-kube-api-access-6xj76\") on node \"crc\" DevicePath \"\"" Feb 14 12:12:48 crc kubenswrapper[4904]: I0214 12:12:48.075678 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db7bd378-cb94-4ae1-b252-420f751b87bf-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 12:12:48 crc kubenswrapper[4904]: I0214 12:12:48.465387 4904 generic.go:334] "Generic (PLEG): container finished" podID="db7bd378-cb94-4ae1-b252-420f751b87bf" containerID="7defe3fe51e1cf297682fe42bae8b508e126ad9cd668347760f25f76796dbccd" exitCode=0 Feb 14 12:12:48 crc kubenswrapper[4904]: I0214 12:12:48.465424 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h68jl" event={"ID":"db7bd378-cb94-4ae1-b252-420f751b87bf","Type":"ContainerDied","Data":"7defe3fe51e1cf297682fe42bae8b508e126ad9cd668347760f25f76796dbccd"} Feb 14 12:12:48 crc kubenswrapper[4904]: I0214 12:12:48.465464 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h68jl" event={"ID":"db7bd378-cb94-4ae1-b252-420f751b87bf","Type":"ContainerDied","Data":"52f1fa345169d2b5856761cbd126b640c61c3dc7fe49af22bb13c377a489be8c"} Feb 14 12:12:48 crc kubenswrapper[4904]: I0214 12:12:48.465483 4904 scope.go:117] "RemoveContainer" containerID="7defe3fe51e1cf297682fe42bae8b508e126ad9cd668347760f25f76796dbccd" Feb 14 12:12:48 crc kubenswrapper[4904]: I0214 12:12:48.466645 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h68jl" Feb 14 12:12:48 crc kubenswrapper[4904]: I0214 12:12:48.489980 4904 scope.go:117] "RemoveContainer" containerID="e57122af2d923f7eea52c61db535b0cca385e0f6435f35136aca86e376a5c09a" Feb 14 12:12:48 crc kubenswrapper[4904]: I0214 12:12:48.504608 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h68jl"] Feb 14 12:12:48 crc kubenswrapper[4904]: I0214 12:12:48.527150 4904 scope.go:117] "RemoveContainer" containerID="b47818f45fe7992f792e971d5c9ca786d1571c1207c7d7012fd2e4aa258b2853" Feb 14 12:12:48 crc kubenswrapper[4904]: I0214 12:12:48.527279 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h68jl"] Feb 14 12:12:48 crc kubenswrapper[4904]: I0214 12:12:48.565022 4904 scope.go:117] "RemoveContainer" containerID="7defe3fe51e1cf297682fe42bae8b508e126ad9cd668347760f25f76796dbccd" Feb 14 12:12:48 crc kubenswrapper[4904]: E0214 12:12:48.565654 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7defe3fe51e1cf297682fe42bae8b508e126ad9cd668347760f25f76796dbccd\": container with ID starting with 7defe3fe51e1cf297682fe42bae8b508e126ad9cd668347760f25f76796dbccd not found: ID does not exist" containerID="7defe3fe51e1cf297682fe42bae8b508e126ad9cd668347760f25f76796dbccd" Feb 14 12:12:48 crc kubenswrapper[4904]: I0214 12:12:48.565726 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7defe3fe51e1cf297682fe42bae8b508e126ad9cd668347760f25f76796dbccd"} err="failed to get container status \"7defe3fe51e1cf297682fe42bae8b508e126ad9cd668347760f25f76796dbccd\": rpc error: code = NotFound desc = could not find container \"7defe3fe51e1cf297682fe42bae8b508e126ad9cd668347760f25f76796dbccd\": container with ID starting with 7defe3fe51e1cf297682fe42bae8b508e126ad9cd668347760f25f76796dbccd not found: ID does not exist" Feb 14 12:12:48 crc kubenswrapper[4904]: I0214 12:12:48.565786 4904 scope.go:117] "RemoveContainer" containerID="e57122af2d923f7eea52c61db535b0cca385e0f6435f35136aca86e376a5c09a" Feb 14 12:12:48 crc kubenswrapper[4904]: E0214 12:12:48.566293 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e57122af2d923f7eea52c61db535b0cca385e0f6435f35136aca86e376a5c09a\": container with ID starting with e57122af2d923f7eea52c61db535b0cca385e0f6435f35136aca86e376a5c09a not found: ID does not exist" containerID="e57122af2d923f7eea52c61db535b0cca385e0f6435f35136aca86e376a5c09a" Feb 14 12:12:48 crc kubenswrapper[4904]: I0214 12:12:48.566323 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e57122af2d923f7eea52c61db535b0cca385e0f6435f35136aca86e376a5c09a"} err="failed to get container status \"e57122af2d923f7eea52c61db535b0cca385e0f6435f35136aca86e376a5c09a\": rpc error: code = NotFound desc = could not find container \"e57122af2d923f7eea52c61db535b0cca385e0f6435f35136aca86e376a5c09a\": container with ID starting with e57122af2d923f7eea52c61db535b0cca385e0f6435f35136aca86e376a5c09a not found: ID does not exist" Feb 14 12:12:48 crc kubenswrapper[4904]: I0214 12:12:48.566361 4904 scope.go:117] "RemoveContainer" containerID="b47818f45fe7992f792e971d5c9ca786d1571c1207c7d7012fd2e4aa258b2853" Feb 14 12:12:48 crc kubenswrapper[4904]: E0214 12:12:48.566687 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b47818f45fe7992f792e971d5c9ca786d1571c1207c7d7012fd2e4aa258b2853\": container with ID starting with b47818f45fe7992f792e971d5c9ca786d1571c1207c7d7012fd2e4aa258b2853 not found: ID does not exist" containerID="b47818f45fe7992f792e971d5c9ca786d1571c1207c7d7012fd2e4aa258b2853" Feb 14 12:12:48 crc kubenswrapper[4904]: I0214 12:12:48.566717 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b47818f45fe7992f792e971d5c9ca786d1571c1207c7d7012fd2e4aa258b2853"} err="failed to get container status \"b47818f45fe7992f792e971d5c9ca786d1571c1207c7d7012fd2e4aa258b2853\": rpc error: code = NotFound desc = could not find container \"b47818f45fe7992f792e971d5c9ca786d1571c1207c7d7012fd2e4aa258b2853\": container with ID starting with b47818f45fe7992f792e971d5c9ca786d1571c1207c7d7012fd2e4aa258b2853 not found: ID does not exist" Feb 14 12:12:49 crc kubenswrapper[4904]: I0214 12:12:49.855535 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db7bd378-cb94-4ae1-b252-420f751b87bf" path="/var/lib/kubelet/pods/db7bd378-cb94-4ae1-b252-420f751b87bf/volumes" Feb 14 12:12:54 crc kubenswrapper[4904]: I0214 12:12:53.838584 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:12:54 crc kubenswrapper[4904]: E0214 12:12:53.839167 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:13:06 crc kubenswrapper[4904]: I0214 12:13:06.836697 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:13:06 crc kubenswrapper[4904]: E0214 12:13:06.837368 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:13:17 crc kubenswrapper[4904]: I0214 12:13:17.836722 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:13:17 crc kubenswrapper[4904]: E0214 12:13:17.838224 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:13:19 crc kubenswrapper[4904]: I0214 12:13:19.510067 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zfmjj"] Feb 14 12:13:19 crc kubenswrapper[4904]: E0214 12:13:19.511819 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7bd378-cb94-4ae1-b252-420f751b87bf" containerName="extract-content" Feb 14 12:13:19 crc kubenswrapper[4904]: I0214 12:13:19.511948 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7bd378-cb94-4ae1-b252-420f751b87bf" containerName="extract-content" Feb 14 12:13:19 crc kubenswrapper[4904]: E0214 12:13:19.512033 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7bd378-cb94-4ae1-b252-420f751b87bf" containerName="registry-server" Feb 14 12:13:19 crc kubenswrapper[4904]: I0214 12:13:19.512111 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7bd378-cb94-4ae1-b252-420f751b87bf" containerName="registry-server" Feb 14 12:13:19 crc kubenswrapper[4904]: E0214 12:13:19.512196 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7bd378-cb94-4ae1-b252-420f751b87bf" containerName="extract-utilities" Feb 14 12:13:19 crc kubenswrapper[4904]: I0214 12:13:19.512274 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7bd378-cb94-4ae1-b252-420f751b87bf" containerName="extract-utilities" Feb 14 12:13:19 crc kubenswrapper[4904]: I0214 12:13:19.512567 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="db7bd378-cb94-4ae1-b252-420f751b87bf" containerName="registry-server" Feb 14 12:13:19 crc kubenswrapper[4904]: I0214 12:13:19.514308 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfmjj" Feb 14 12:13:19 crc kubenswrapper[4904]: I0214 12:13:19.521309 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zfmjj"] Feb 14 12:13:19 crc kubenswrapper[4904]: I0214 12:13:19.614315 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsk9l\" (UniqueName: \"kubernetes.io/projected/29b175e0-948f-43aa-b3b2-91224558045f-kube-api-access-gsk9l\") pod \"community-operators-zfmjj\" (UID: \"29b175e0-948f-43aa-b3b2-91224558045f\") " pod="openshift-marketplace/community-operators-zfmjj" Feb 14 12:13:19 crc kubenswrapper[4904]: I0214 12:13:19.614415 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29b175e0-948f-43aa-b3b2-91224558045f-catalog-content\") pod \"community-operators-zfmjj\" (UID: \"29b175e0-948f-43aa-b3b2-91224558045f\") " pod="openshift-marketplace/community-operators-zfmjj" Feb 14 12:13:19 crc kubenswrapper[4904]: I0214 12:13:19.614444 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29b175e0-948f-43aa-b3b2-91224558045f-utilities\") pod \"community-operators-zfmjj\" (UID: \"29b175e0-948f-43aa-b3b2-91224558045f\") " pod="openshift-marketplace/community-operators-zfmjj" Feb 14 12:13:19 crc kubenswrapper[4904]: I0214 12:13:19.715950 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsk9l\" (UniqueName: \"kubernetes.io/projected/29b175e0-948f-43aa-b3b2-91224558045f-kube-api-access-gsk9l\") pod \"community-operators-zfmjj\" (UID: \"29b175e0-948f-43aa-b3b2-91224558045f\") " pod="openshift-marketplace/community-operators-zfmjj" Feb 14 12:13:19 crc kubenswrapper[4904]: I0214 12:13:19.716031 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29b175e0-948f-43aa-b3b2-91224558045f-catalog-content\") pod \"community-operators-zfmjj\" (UID: \"29b175e0-948f-43aa-b3b2-91224558045f\") " pod="openshift-marketplace/community-operators-zfmjj" Feb 14 12:13:19 crc kubenswrapper[4904]: I0214 12:13:19.716055 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29b175e0-948f-43aa-b3b2-91224558045f-utilities\") pod \"community-operators-zfmjj\" (UID: \"29b175e0-948f-43aa-b3b2-91224558045f\") " pod="openshift-marketplace/community-operators-zfmjj" Feb 14 12:13:19 crc kubenswrapper[4904]: I0214 12:13:19.716571 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29b175e0-948f-43aa-b3b2-91224558045f-utilities\") pod \"community-operators-zfmjj\" (UID: \"29b175e0-948f-43aa-b3b2-91224558045f\") " pod="openshift-marketplace/community-operators-zfmjj" Feb 14 12:13:19 crc kubenswrapper[4904]: I0214 12:13:19.716694 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29b175e0-948f-43aa-b3b2-91224558045f-catalog-content\") pod \"community-operators-zfmjj\" (UID: \"29b175e0-948f-43aa-b3b2-91224558045f\") " pod="openshift-marketplace/community-operators-zfmjj" Feb 14 12:13:19 crc kubenswrapper[4904]: I0214 12:13:19.737465 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsk9l\" (UniqueName: \"kubernetes.io/projected/29b175e0-948f-43aa-b3b2-91224558045f-kube-api-access-gsk9l\") pod \"community-operators-zfmjj\" (UID: \"29b175e0-948f-43aa-b3b2-91224558045f\") " pod="openshift-marketplace/community-operators-zfmjj" Feb 14 12:13:19 crc kubenswrapper[4904]: I0214 12:13:19.836115 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfmjj" Feb 14 12:13:20 crc kubenswrapper[4904]: I0214 12:13:20.458298 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zfmjj"] Feb 14 12:13:20 crc kubenswrapper[4904]: I0214 12:13:20.723909 4904 generic.go:334] "Generic (PLEG): container finished" podID="29b175e0-948f-43aa-b3b2-91224558045f" containerID="e1000dec81dc16924c2e5ee4f75ca0c6d4a4061c8c5cb219b2352d54aacd6415" exitCode=0 Feb 14 12:13:20 crc kubenswrapper[4904]: I0214 12:13:20.723951 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfmjj" event={"ID":"29b175e0-948f-43aa-b3b2-91224558045f","Type":"ContainerDied","Data":"e1000dec81dc16924c2e5ee4f75ca0c6d4a4061c8c5cb219b2352d54aacd6415"} Feb 14 12:13:20 crc kubenswrapper[4904]: I0214 12:13:20.723976 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfmjj" event={"ID":"29b175e0-948f-43aa-b3b2-91224558045f","Type":"ContainerStarted","Data":"f902b853f40811a6d155a2c21354b49594c934bd461f48750910d6299061f10d"} Feb 14 12:13:20 crc kubenswrapper[4904]: I0214 12:13:20.726501 4904 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 14 12:13:21 crc kubenswrapper[4904]: I0214 12:13:21.734236 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfmjj" event={"ID":"29b175e0-948f-43aa-b3b2-91224558045f","Type":"ContainerStarted","Data":"974abbd9c4e080977d85e3ec271ff80870395b4ff6873ca7172d0fb32eb3d763"} Feb 14 12:13:22 crc kubenswrapper[4904]: I0214 12:13:22.743437 4904 generic.go:334] "Generic (PLEG): container finished" podID="29b175e0-948f-43aa-b3b2-91224558045f" containerID="974abbd9c4e080977d85e3ec271ff80870395b4ff6873ca7172d0fb32eb3d763" exitCode=0 Feb 14 12:13:22 crc kubenswrapper[4904]: I0214 12:13:22.743551 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfmjj" event={"ID":"29b175e0-948f-43aa-b3b2-91224558045f","Type":"ContainerDied","Data":"974abbd9c4e080977d85e3ec271ff80870395b4ff6873ca7172d0fb32eb3d763"} Feb 14 12:13:23 crc kubenswrapper[4904]: I0214 12:13:23.752901 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfmjj" event={"ID":"29b175e0-948f-43aa-b3b2-91224558045f","Type":"ContainerStarted","Data":"96ddca4bf530aeab79586bb4696d9d393e71ac6920ff79eb965ef1ea48ae9413"} Feb 14 12:13:23 crc kubenswrapper[4904]: I0214 12:13:23.775233 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zfmjj" podStartSLOduration=2.358795578 podStartE2EDuration="4.775211163s" podCreationTimestamp="2026-02-14 12:13:19 +0000 UTC" firstStartedPulling="2026-02-14 12:13:20.726232231 +0000 UTC m=+3791.538996892" lastFinishedPulling="2026-02-14 12:13:23.142647816 +0000 UTC m=+3793.955412477" observedRunningTime="2026-02-14 12:13:23.769518737 +0000 UTC m=+3794.582283398" watchObservedRunningTime="2026-02-14 12:13:23.775211163 +0000 UTC m=+3794.587975824" Feb 14 12:13:29 crc kubenswrapper[4904]: I0214 12:13:29.856776 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zfmjj" Feb 14 12:13:29 crc kubenswrapper[4904]: I0214 12:13:29.857634 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zfmjj" Feb 14 12:13:29 crc kubenswrapper[4904]: I0214 12:13:29.917150 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zfmjj" Feb 14 12:13:30 crc kubenswrapper[4904]: I0214 12:13:30.836085 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:13:30 crc kubenswrapper[4904]: E0214 12:13:30.836534 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:13:30 crc kubenswrapper[4904]: I0214 12:13:30.873181 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zfmjj" Feb 14 12:13:30 crc kubenswrapper[4904]: I0214 12:13:30.920682 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zfmjj"] Feb 14 12:13:32 crc kubenswrapper[4904]: I0214 12:13:32.852076 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zfmjj" podUID="29b175e0-948f-43aa-b3b2-91224558045f" containerName="registry-server" containerID="cri-o://96ddca4bf530aeab79586bb4696d9d393e71ac6920ff79eb965ef1ea48ae9413" gracePeriod=2 Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.259037 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfmjj" Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.369589 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29b175e0-948f-43aa-b3b2-91224558045f-utilities\") pod \"29b175e0-948f-43aa-b3b2-91224558045f\" (UID: \"29b175e0-948f-43aa-b3b2-91224558045f\") " Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.369671 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsk9l\" (UniqueName: \"kubernetes.io/projected/29b175e0-948f-43aa-b3b2-91224558045f-kube-api-access-gsk9l\") pod \"29b175e0-948f-43aa-b3b2-91224558045f\" (UID: \"29b175e0-948f-43aa-b3b2-91224558045f\") " Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.369716 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29b175e0-948f-43aa-b3b2-91224558045f-catalog-content\") pod \"29b175e0-948f-43aa-b3b2-91224558045f\" (UID: \"29b175e0-948f-43aa-b3b2-91224558045f\") " Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.371817 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29b175e0-948f-43aa-b3b2-91224558045f-utilities" (OuterVolumeSpecName: "utilities") pod "29b175e0-948f-43aa-b3b2-91224558045f" (UID: "29b175e0-948f-43aa-b3b2-91224558045f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.378534 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29b175e0-948f-43aa-b3b2-91224558045f-kube-api-access-gsk9l" (OuterVolumeSpecName: "kube-api-access-gsk9l") pod "29b175e0-948f-43aa-b3b2-91224558045f" (UID: "29b175e0-948f-43aa-b3b2-91224558045f"). InnerVolumeSpecName "kube-api-access-gsk9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.420762 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29b175e0-948f-43aa-b3b2-91224558045f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "29b175e0-948f-43aa-b3b2-91224558045f" (UID: "29b175e0-948f-43aa-b3b2-91224558045f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.472809 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29b175e0-948f-43aa-b3b2-91224558045f-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.473451 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsk9l\" (UniqueName: \"kubernetes.io/projected/29b175e0-948f-43aa-b3b2-91224558045f-kube-api-access-gsk9l\") on node \"crc\" DevicePath \"\"" Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.473474 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29b175e0-948f-43aa-b3b2-91224558045f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.860205 4904 generic.go:334] "Generic (PLEG): container finished" podID="29b175e0-948f-43aa-b3b2-91224558045f" containerID="96ddca4bf530aeab79586bb4696d9d393e71ac6920ff79eb965ef1ea48ae9413" exitCode=0 Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.860251 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfmjj" event={"ID":"29b175e0-948f-43aa-b3b2-91224558045f","Type":"ContainerDied","Data":"96ddca4bf530aeab79586bb4696d9d393e71ac6920ff79eb965ef1ea48ae9413"} Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.860287 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfmjj" event={"ID":"29b175e0-948f-43aa-b3b2-91224558045f","Type":"ContainerDied","Data":"f902b853f40811a6d155a2c21354b49594c934bd461f48750910d6299061f10d"} Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.860305 4904 scope.go:117] "RemoveContainer" containerID="96ddca4bf530aeab79586bb4696d9d393e71ac6920ff79eb965ef1ea48ae9413" Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.860317 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfmjj" Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.888016 4904 scope.go:117] "RemoveContainer" containerID="974abbd9c4e080977d85e3ec271ff80870395b4ff6873ca7172d0fb32eb3d763" Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.888710 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zfmjj"] Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.902026 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zfmjj"] Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.914522 4904 scope.go:117] "RemoveContainer" containerID="e1000dec81dc16924c2e5ee4f75ca0c6d4a4061c8c5cb219b2352d54aacd6415" Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.955876 4904 scope.go:117] "RemoveContainer" containerID="96ddca4bf530aeab79586bb4696d9d393e71ac6920ff79eb965ef1ea48ae9413" Feb 14 12:13:33 crc kubenswrapper[4904]: E0214 12:13:33.956243 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96ddca4bf530aeab79586bb4696d9d393e71ac6920ff79eb965ef1ea48ae9413\": container with ID starting with 96ddca4bf530aeab79586bb4696d9d393e71ac6920ff79eb965ef1ea48ae9413 not found: ID does not exist" containerID="96ddca4bf530aeab79586bb4696d9d393e71ac6920ff79eb965ef1ea48ae9413" Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.956289 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96ddca4bf530aeab79586bb4696d9d393e71ac6920ff79eb965ef1ea48ae9413"} err="failed to get container status \"96ddca4bf530aeab79586bb4696d9d393e71ac6920ff79eb965ef1ea48ae9413\": rpc error: code = NotFound desc = could not find container \"96ddca4bf530aeab79586bb4696d9d393e71ac6920ff79eb965ef1ea48ae9413\": container with ID starting with 96ddca4bf530aeab79586bb4696d9d393e71ac6920ff79eb965ef1ea48ae9413 not found: ID does not exist" Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.956316 4904 scope.go:117] "RemoveContainer" containerID="974abbd9c4e080977d85e3ec271ff80870395b4ff6873ca7172d0fb32eb3d763" Feb 14 12:13:33 crc kubenswrapper[4904]: E0214 12:13:33.956650 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"974abbd9c4e080977d85e3ec271ff80870395b4ff6873ca7172d0fb32eb3d763\": container with ID starting with 974abbd9c4e080977d85e3ec271ff80870395b4ff6873ca7172d0fb32eb3d763 not found: ID does not exist" containerID="974abbd9c4e080977d85e3ec271ff80870395b4ff6873ca7172d0fb32eb3d763" Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.956686 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"974abbd9c4e080977d85e3ec271ff80870395b4ff6873ca7172d0fb32eb3d763"} err="failed to get container status \"974abbd9c4e080977d85e3ec271ff80870395b4ff6873ca7172d0fb32eb3d763\": rpc error: code = NotFound desc = could not find container \"974abbd9c4e080977d85e3ec271ff80870395b4ff6873ca7172d0fb32eb3d763\": container with ID starting with 974abbd9c4e080977d85e3ec271ff80870395b4ff6873ca7172d0fb32eb3d763 not found: ID does not exist" Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.956708 4904 scope.go:117] "RemoveContainer" containerID="e1000dec81dc16924c2e5ee4f75ca0c6d4a4061c8c5cb219b2352d54aacd6415" Feb 14 12:13:33 crc kubenswrapper[4904]: E0214 12:13:33.957036 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1000dec81dc16924c2e5ee4f75ca0c6d4a4061c8c5cb219b2352d54aacd6415\": container with ID starting with e1000dec81dc16924c2e5ee4f75ca0c6d4a4061c8c5cb219b2352d54aacd6415 not found: ID does not exist" containerID="e1000dec81dc16924c2e5ee4f75ca0c6d4a4061c8c5cb219b2352d54aacd6415" Feb 14 12:13:33 crc kubenswrapper[4904]: I0214 12:13:33.957066 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1000dec81dc16924c2e5ee4f75ca0c6d4a4061c8c5cb219b2352d54aacd6415"} err="failed to get container status \"e1000dec81dc16924c2e5ee4f75ca0c6d4a4061c8c5cb219b2352d54aacd6415\": rpc error: code = NotFound desc = could not find container \"e1000dec81dc16924c2e5ee4f75ca0c6d4a4061c8c5cb219b2352d54aacd6415\": container with ID starting with e1000dec81dc16924c2e5ee4f75ca0c6d4a4061c8c5cb219b2352d54aacd6415 not found: ID does not exist" Feb 14 12:13:35 crc kubenswrapper[4904]: I0214 12:13:35.850661 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29b175e0-948f-43aa-b3b2-91224558045f" path="/var/lib/kubelet/pods/29b175e0-948f-43aa-b3b2-91224558045f/volumes" Feb 14 12:13:42 crc kubenswrapper[4904]: I0214 12:13:42.736440 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-668mk"] Feb 14 12:13:42 crc kubenswrapper[4904]: E0214 12:13:42.741373 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29b175e0-948f-43aa-b3b2-91224558045f" containerName="extract-utilities" Feb 14 12:13:42 crc kubenswrapper[4904]: I0214 12:13:42.741401 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="29b175e0-948f-43aa-b3b2-91224558045f" containerName="extract-utilities" Feb 14 12:13:42 crc kubenswrapper[4904]: E0214 12:13:42.741431 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29b175e0-948f-43aa-b3b2-91224558045f" containerName="registry-server" Feb 14 12:13:42 crc kubenswrapper[4904]: I0214 12:13:42.741440 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="29b175e0-948f-43aa-b3b2-91224558045f" containerName="registry-server" Feb 14 12:13:42 crc kubenswrapper[4904]: E0214 12:13:42.741458 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29b175e0-948f-43aa-b3b2-91224558045f" containerName="extract-content" Feb 14 12:13:42 crc kubenswrapper[4904]: I0214 12:13:42.741466 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="29b175e0-948f-43aa-b3b2-91224558045f" containerName="extract-content" Feb 14 12:13:42 crc kubenswrapper[4904]: I0214 12:13:42.741655 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="29b175e0-948f-43aa-b3b2-91224558045f" containerName="registry-server" Feb 14 12:13:42 crc kubenswrapper[4904]: I0214 12:13:42.743272 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-668mk" Feb 14 12:13:42 crc kubenswrapper[4904]: I0214 12:13:42.746925 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2knf\" (UniqueName: \"kubernetes.io/projected/b2ae5940-3808-4edd-8618-06f4274ad4aa-kube-api-access-t2knf\") pod \"certified-operators-668mk\" (UID: \"b2ae5940-3808-4edd-8618-06f4274ad4aa\") " pod="openshift-marketplace/certified-operators-668mk" Feb 14 12:13:42 crc kubenswrapper[4904]: I0214 12:13:42.746972 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ae5940-3808-4edd-8618-06f4274ad4aa-catalog-content\") pod \"certified-operators-668mk\" (UID: \"b2ae5940-3808-4edd-8618-06f4274ad4aa\") " pod="openshift-marketplace/certified-operators-668mk" Feb 14 12:13:42 crc kubenswrapper[4904]: I0214 12:13:42.747121 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ae5940-3808-4edd-8618-06f4274ad4aa-utilities\") pod \"certified-operators-668mk\" (UID: \"b2ae5940-3808-4edd-8618-06f4274ad4aa\") " pod="openshift-marketplace/certified-operators-668mk" Feb 14 12:13:42 crc kubenswrapper[4904]: I0214 12:13:42.768097 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-668mk"] Feb 14 12:13:42 crc kubenswrapper[4904]: I0214 12:13:42.848989 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2knf\" (UniqueName: \"kubernetes.io/projected/b2ae5940-3808-4edd-8618-06f4274ad4aa-kube-api-access-t2knf\") pod \"certified-operators-668mk\" (UID: \"b2ae5940-3808-4edd-8618-06f4274ad4aa\") " pod="openshift-marketplace/certified-operators-668mk" Feb 14 12:13:42 crc kubenswrapper[4904]: I0214 12:13:42.849042 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ae5940-3808-4edd-8618-06f4274ad4aa-catalog-content\") pod \"certified-operators-668mk\" (UID: \"b2ae5940-3808-4edd-8618-06f4274ad4aa\") " pod="openshift-marketplace/certified-operators-668mk" Feb 14 12:13:42 crc kubenswrapper[4904]: I0214 12:13:42.849151 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ae5940-3808-4edd-8618-06f4274ad4aa-utilities\") pod \"certified-operators-668mk\" (UID: \"b2ae5940-3808-4edd-8618-06f4274ad4aa\") " pod="openshift-marketplace/certified-operators-668mk" Feb 14 12:13:42 crc kubenswrapper[4904]: I0214 12:13:42.849543 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ae5940-3808-4edd-8618-06f4274ad4aa-catalog-content\") pod \"certified-operators-668mk\" (UID: \"b2ae5940-3808-4edd-8618-06f4274ad4aa\") " pod="openshift-marketplace/certified-operators-668mk" Feb 14 12:13:42 crc kubenswrapper[4904]: I0214 12:13:42.849943 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ae5940-3808-4edd-8618-06f4274ad4aa-utilities\") pod \"certified-operators-668mk\" (UID: \"b2ae5940-3808-4edd-8618-06f4274ad4aa\") " pod="openshift-marketplace/certified-operators-668mk" Feb 14 12:13:42 crc kubenswrapper[4904]: I0214 12:13:42.876347 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2knf\" (UniqueName: \"kubernetes.io/projected/b2ae5940-3808-4edd-8618-06f4274ad4aa-kube-api-access-t2knf\") pod \"certified-operators-668mk\" (UID: \"b2ae5940-3808-4edd-8618-06f4274ad4aa\") " pod="openshift-marketplace/certified-operators-668mk" Feb 14 12:13:43 crc kubenswrapper[4904]: I0214 12:13:43.059668 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-668mk" Feb 14 12:13:43 crc kubenswrapper[4904]: I0214 12:13:43.630080 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-668mk"] Feb 14 12:13:43 crc kubenswrapper[4904]: I0214 12:13:43.937910 4904 generic.go:334] "Generic (PLEG): container finished" podID="b2ae5940-3808-4edd-8618-06f4274ad4aa" containerID="e9a27d3d3ad80348b7a309942e5342550bbb26974165133685111706680a39a9" exitCode=0 Feb 14 12:13:43 crc kubenswrapper[4904]: I0214 12:13:43.937959 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-668mk" event={"ID":"b2ae5940-3808-4edd-8618-06f4274ad4aa","Type":"ContainerDied","Data":"e9a27d3d3ad80348b7a309942e5342550bbb26974165133685111706680a39a9"} Feb 14 12:13:43 crc kubenswrapper[4904]: I0214 12:13:43.937989 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-668mk" event={"ID":"b2ae5940-3808-4edd-8618-06f4274ad4aa","Type":"ContainerStarted","Data":"803f14e8770d305e3e2e05e513f9612e7a872ad21e3d8691329d0797a79c222c"} Feb 14 12:13:44 crc kubenswrapper[4904]: I0214 12:13:44.946795 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-668mk" event={"ID":"b2ae5940-3808-4edd-8618-06f4274ad4aa","Type":"ContainerStarted","Data":"bf7e1da9192364a0af449c03839c4605f382e035fc04f795c557e6352c7eb06f"} Feb 14 12:13:45 crc kubenswrapper[4904]: I0214 12:13:45.838503 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:13:45 crc kubenswrapper[4904]: E0214 12:13:45.839358 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:13:46 crc kubenswrapper[4904]: I0214 12:13:46.962186 4904 generic.go:334] "Generic (PLEG): container finished" podID="b2ae5940-3808-4edd-8618-06f4274ad4aa" containerID="bf7e1da9192364a0af449c03839c4605f382e035fc04f795c557e6352c7eb06f" exitCode=0 Feb 14 12:13:46 crc kubenswrapper[4904]: I0214 12:13:46.962228 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-668mk" event={"ID":"b2ae5940-3808-4edd-8618-06f4274ad4aa","Type":"ContainerDied","Data":"bf7e1da9192364a0af449c03839c4605f382e035fc04f795c557e6352c7eb06f"} Feb 14 12:13:47 crc kubenswrapper[4904]: I0214 12:13:47.980299 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-668mk" event={"ID":"b2ae5940-3808-4edd-8618-06f4274ad4aa","Type":"ContainerStarted","Data":"842778ee2b83b9503cb298539716138d4cb7856da9339a04e2ad45ba1b8a9e8c"} Feb 14 12:13:48 crc kubenswrapper[4904]: I0214 12:13:48.009279 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-668mk" podStartSLOduration=2.574479013 podStartE2EDuration="6.009260413s" podCreationTimestamp="2026-02-14 12:13:42 +0000 UTC" firstStartedPulling="2026-02-14 12:13:43.939565381 +0000 UTC m=+3814.752330042" lastFinishedPulling="2026-02-14 12:13:47.374346781 +0000 UTC m=+3818.187111442" observedRunningTime="2026-02-14 12:13:47.998761145 +0000 UTC m=+3818.811525806" watchObservedRunningTime="2026-02-14 12:13:48.009260413 +0000 UTC m=+3818.822025074" Feb 14 12:13:53 crc kubenswrapper[4904]: I0214 12:13:53.060377 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-668mk" Feb 14 12:13:53 crc kubenswrapper[4904]: I0214 12:13:53.061230 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-668mk" Feb 14 12:13:53 crc kubenswrapper[4904]: I0214 12:13:53.121655 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-668mk" Feb 14 12:13:54 crc kubenswrapper[4904]: I0214 12:13:54.110262 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-668mk" Feb 14 12:13:54 crc kubenswrapper[4904]: I0214 12:13:54.174331 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-668mk"] Feb 14 12:13:56 crc kubenswrapper[4904]: I0214 12:13:56.061095 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-668mk" podUID="b2ae5940-3808-4edd-8618-06f4274ad4aa" containerName="registry-server" containerID="cri-o://842778ee2b83b9503cb298539716138d4cb7856da9339a04e2ad45ba1b8a9e8c" gracePeriod=2 Feb 14 12:13:56 crc kubenswrapper[4904]: I0214 12:13:56.517967 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-668mk" Feb 14 12:13:56 crc kubenswrapper[4904]: I0214 12:13:56.619822 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ae5940-3808-4edd-8618-06f4274ad4aa-catalog-content\") pod \"b2ae5940-3808-4edd-8618-06f4274ad4aa\" (UID: \"b2ae5940-3808-4edd-8618-06f4274ad4aa\") " Feb 14 12:13:56 crc kubenswrapper[4904]: I0214 12:13:56.619944 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2knf\" (UniqueName: \"kubernetes.io/projected/b2ae5940-3808-4edd-8618-06f4274ad4aa-kube-api-access-t2knf\") pod \"b2ae5940-3808-4edd-8618-06f4274ad4aa\" (UID: \"b2ae5940-3808-4edd-8618-06f4274ad4aa\") " Feb 14 12:13:56 crc kubenswrapper[4904]: I0214 12:13:56.620066 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ae5940-3808-4edd-8618-06f4274ad4aa-utilities\") pod \"b2ae5940-3808-4edd-8618-06f4274ad4aa\" (UID: \"b2ae5940-3808-4edd-8618-06f4274ad4aa\") " Feb 14 12:13:56 crc kubenswrapper[4904]: I0214 12:13:56.621636 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2ae5940-3808-4edd-8618-06f4274ad4aa-utilities" (OuterVolumeSpecName: "utilities") pod "b2ae5940-3808-4edd-8618-06f4274ad4aa" (UID: "b2ae5940-3808-4edd-8618-06f4274ad4aa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:13:56 crc kubenswrapper[4904]: I0214 12:13:56.625305 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2ae5940-3808-4edd-8618-06f4274ad4aa-kube-api-access-t2knf" (OuterVolumeSpecName: "kube-api-access-t2knf") pod "b2ae5940-3808-4edd-8618-06f4274ad4aa" (UID: "b2ae5940-3808-4edd-8618-06f4274ad4aa"). InnerVolumeSpecName "kube-api-access-t2knf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:13:56 crc kubenswrapper[4904]: I0214 12:13:56.674154 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2ae5940-3808-4edd-8618-06f4274ad4aa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2ae5940-3808-4edd-8618-06f4274ad4aa" (UID: "b2ae5940-3808-4edd-8618-06f4274ad4aa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:13:56 crc kubenswrapper[4904]: I0214 12:13:56.723388 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ae5940-3808-4edd-8618-06f4274ad4aa-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 12:13:56 crc kubenswrapper[4904]: I0214 12:13:56.723506 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ae5940-3808-4edd-8618-06f4274ad4aa-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 12:13:56 crc kubenswrapper[4904]: I0214 12:13:56.723543 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2knf\" (UniqueName: \"kubernetes.io/projected/b2ae5940-3808-4edd-8618-06f4274ad4aa-kube-api-access-t2knf\") on node \"crc\" DevicePath \"\"" Feb 14 12:13:57 crc kubenswrapper[4904]: I0214 12:13:57.069159 4904 generic.go:334] "Generic (PLEG): container finished" podID="b2ae5940-3808-4edd-8618-06f4274ad4aa" containerID="842778ee2b83b9503cb298539716138d4cb7856da9339a04e2ad45ba1b8a9e8c" exitCode=0 Feb 14 12:13:57 crc kubenswrapper[4904]: I0214 12:13:57.069203 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-668mk" event={"ID":"b2ae5940-3808-4edd-8618-06f4274ad4aa","Type":"ContainerDied","Data":"842778ee2b83b9503cb298539716138d4cb7856da9339a04e2ad45ba1b8a9e8c"} Feb 14 12:13:57 crc kubenswrapper[4904]: I0214 12:13:57.069216 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-668mk" Feb 14 12:13:57 crc kubenswrapper[4904]: I0214 12:13:57.069241 4904 scope.go:117] "RemoveContainer" containerID="842778ee2b83b9503cb298539716138d4cb7856da9339a04e2ad45ba1b8a9e8c" Feb 14 12:13:57 crc kubenswrapper[4904]: I0214 12:13:57.069230 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-668mk" event={"ID":"b2ae5940-3808-4edd-8618-06f4274ad4aa","Type":"ContainerDied","Data":"803f14e8770d305e3e2e05e513f9612e7a872ad21e3d8691329d0797a79c222c"} Feb 14 12:13:57 crc kubenswrapper[4904]: I0214 12:13:57.088662 4904 scope.go:117] "RemoveContainer" containerID="bf7e1da9192364a0af449c03839c4605f382e035fc04f795c557e6352c7eb06f" Feb 14 12:13:57 crc kubenswrapper[4904]: I0214 12:13:57.106945 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-668mk"] Feb 14 12:13:57 crc kubenswrapper[4904]: I0214 12:13:57.117428 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-668mk"] Feb 14 12:13:57 crc kubenswrapper[4904]: I0214 12:13:57.136123 4904 scope.go:117] "RemoveContainer" containerID="e9a27d3d3ad80348b7a309942e5342550bbb26974165133685111706680a39a9" Feb 14 12:13:57 crc kubenswrapper[4904]: I0214 12:13:57.163218 4904 scope.go:117] "RemoveContainer" containerID="842778ee2b83b9503cb298539716138d4cb7856da9339a04e2ad45ba1b8a9e8c" Feb 14 12:13:57 crc kubenswrapper[4904]: E0214 12:13:57.163646 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"842778ee2b83b9503cb298539716138d4cb7856da9339a04e2ad45ba1b8a9e8c\": container with ID starting with 842778ee2b83b9503cb298539716138d4cb7856da9339a04e2ad45ba1b8a9e8c not found: ID does not exist" containerID="842778ee2b83b9503cb298539716138d4cb7856da9339a04e2ad45ba1b8a9e8c" Feb 14 12:13:57 crc kubenswrapper[4904]: I0214 12:13:57.163688 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"842778ee2b83b9503cb298539716138d4cb7856da9339a04e2ad45ba1b8a9e8c"} err="failed to get container status \"842778ee2b83b9503cb298539716138d4cb7856da9339a04e2ad45ba1b8a9e8c\": rpc error: code = NotFound desc = could not find container \"842778ee2b83b9503cb298539716138d4cb7856da9339a04e2ad45ba1b8a9e8c\": container with ID starting with 842778ee2b83b9503cb298539716138d4cb7856da9339a04e2ad45ba1b8a9e8c not found: ID does not exist" Feb 14 12:13:57 crc kubenswrapper[4904]: I0214 12:13:57.163717 4904 scope.go:117] "RemoveContainer" containerID="bf7e1da9192364a0af449c03839c4605f382e035fc04f795c557e6352c7eb06f" Feb 14 12:13:57 crc kubenswrapper[4904]: E0214 12:13:57.164200 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf7e1da9192364a0af449c03839c4605f382e035fc04f795c557e6352c7eb06f\": container with ID starting with bf7e1da9192364a0af449c03839c4605f382e035fc04f795c557e6352c7eb06f not found: ID does not exist" containerID="bf7e1da9192364a0af449c03839c4605f382e035fc04f795c557e6352c7eb06f" Feb 14 12:13:57 crc kubenswrapper[4904]: I0214 12:13:57.164234 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf7e1da9192364a0af449c03839c4605f382e035fc04f795c557e6352c7eb06f"} err="failed to get container status \"bf7e1da9192364a0af449c03839c4605f382e035fc04f795c557e6352c7eb06f\": rpc error: code = NotFound desc = could not find container \"bf7e1da9192364a0af449c03839c4605f382e035fc04f795c557e6352c7eb06f\": container with ID starting with bf7e1da9192364a0af449c03839c4605f382e035fc04f795c557e6352c7eb06f not found: ID does not exist" Feb 14 12:13:57 crc kubenswrapper[4904]: I0214 12:13:57.164259 4904 scope.go:117] "RemoveContainer" containerID="e9a27d3d3ad80348b7a309942e5342550bbb26974165133685111706680a39a9" Feb 14 12:13:57 crc kubenswrapper[4904]: E0214 12:13:57.164475 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9a27d3d3ad80348b7a309942e5342550bbb26974165133685111706680a39a9\": container with ID starting with e9a27d3d3ad80348b7a309942e5342550bbb26974165133685111706680a39a9 not found: ID does not exist" containerID="e9a27d3d3ad80348b7a309942e5342550bbb26974165133685111706680a39a9" Feb 14 12:13:57 crc kubenswrapper[4904]: I0214 12:13:57.164496 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9a27d3d3ad80348b7a309942e5342550bbb26974165133685111706680a39a9"} err="failed to get container status \"e9a27d3d3ad80348b7a309942e5342550bbb26974165133685111706680a39a9\": rpc error: code = NotFound desc = could not find container \"e9a27d3d3ad80348b7a309942e5342550bbb26974165133685111706680a39a9\": container with ID starting with e9a27d3d3ad80348b7a309942e5342550bbb26974165133685111706680a39a9 not found: ID does not exist" Feb 14 12:13:57 crc kubenswrapper[4904]: I0214 12:13:57.836504 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:13:57 crc kubenswrapper[4904]: E0214 12:13:57.836901 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:13:57 crc kubenswrapper[4904]: I0214 12:13:57.854566 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2ae5940-3808-4edd-8618-06f4274ad4aa" path="/var/lib/kubelet/pods/b2ae5940-3808-4edd-8618-06f4274ad4aa/volumes" Feb 14 12:14:12 crc kubenswrapper[4904]: I0214 12:14:12.836139 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:14:12 crc kubenswrapper[4904]: E0214 12:14:12.836804 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:14:24 crc kubenswrapper[4904]: I0214 12:14:24.836965 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:14:24 crc kubenswrapper[4904]: E0214 12:14:24.837788 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:14:36 crc kubenswrapper[4904]: I0214 12:14:36.836559 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:14:36 crc kubenswrapper[4904]: E0214 12:14:36.837298 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:14:47 crc kubenswrapper[4904]: I0214 12:14:47.836788 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:14:47 crc kubenswrapper[4904]: E0214 12:14:47.837523 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.183059 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517855-lsnzv"] Feb 14 12:15:00 crc kubenswrapper[4904]: E0214 12:15:00.184146 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ae5940-3808-4edd-8618-06f4274ad4aa" containerName="extract-content" Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.184166 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ae5940-3808-4edd-8618-06f4274ad4aa" containerName="extract-content" Feb 14 12:15:00 crc kubenswrapper[4904]: E0214 12:15:00.184190 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ae5940-3808-4edd-8618-06f4274ad4aa" containerName="registry-server" Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.184199 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ae5940-3808-4edd-8618-06f4274ad4aa" containerName="registry-server" Feb 14 12:15:00 crc kubenswrapper[4904]: E0214 12:15:00.184240 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ae5940-3808-4edd-8618-06f4274ad4aa" containerName="extract-utilities" Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.184250 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ae5940-3808-4edd-8618-06f4274ad4aa" containerName="extract-utilities" Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.184499 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2ae5940-3808-4edd-8618-06f4274ad4aa" containerName="registry-server" Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.185291 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517855-lsnzv" Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.187922 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.190086 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.200963 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517855-lsnzv"] Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.346329 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckjjv\" (UniqueName: \"kubernetes.io/projected/e3558d0e-d510-4267-8937-f4d3cfe5d178-kube-api-access-ckjjv\") pod \"collect-profiles-29517855-lsnzv\" (UID: \"e3558d0e-d510-4267-8937-f4d3cfe5d178\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517855-lsnzv" Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.346469 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e3558d0e-d510-4267-8937-f4d3cfe5d178-secret-volume\") pod \"collect-profiles-29517855-lsnzv\" (UID: \"e3558d0e-d510-4267-8937-f4d3cfe5d178\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517855-lsnzv" Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.346617 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e3558d0e-d510-4267-8937-f4d3cfe5d178-config-volume\") pod \"collect-profiles-29517855-lsnzv\" (UID: \"e3558d0e-d510-4267-8937-f4d3cfe5d178\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517855-lsnzv" Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.448702 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckjjv\" (UniqueName: \"kubernetes.io/projected/e3558d0e-d510-4267-8937-f4d3cfe5d178-kube-api-access-ckjjv\") pod \"collect-profiles-29517855-lsnzv\" (UID: \"e3558d0e-d510-4267-8937-f4d3cfe5d178\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517855-lsnzv" Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.448777 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e3558d0e-d510-4267-8937-f4d3cfe5d178-secret-volume\") pod \"collect-profiles-29517855-lsnzv\" (UID: \"e3558d0e-d510-4267-8937-f4d3cfe5d178\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517855-lsnzv" Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.448944 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e3558d0e-d510-4267-8937-f4d3cfe5d178-config-volume\") pod \"collect-profiles-29517855-lsnzv\" (UID: \"e3558d0e-d510-4267-8937-f4d3cfe5d178\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517855-lsnzv" Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.449858 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e3558d0e-d510-4267-8937-f4d3cfe5d178-config-volume\") pod \"collect-profiles-29517855-lsnzv\" (UID: \"e3558d0e-d510-4267-8937-f4d3cfe5d178\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517855-lsnzv" Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.455093 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e3558d0e-d510-4267-8937-f4d3cfe5d178-secret-volume\") pod \"collect-profiles-29517855-lsnzv\" (UID: \"e3558d0e-d510-4267-8937-f4d3cfe5d178\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517855-lsnzv" Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.467263 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckjjv\" (UniqueName: \"kubernetes.io/projected/e3558d0e-d510-4267-8937-f4d3cfe5d178-kube-api-access-ckjjv\") pod \"collect-profiles-29517855-lsnzv\" (UID: \"e3558d0e-d510-4267-8937-f4d3cfe5d178\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517855-lsnzv" Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.506980 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517855-lsnzv" Feb 14 12:15:00 crc kubenswrapper[4904]: I0214 12:15:00.945950 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517855-lsnzv"] Feb 14 12:15:01 crc kubenswrapper[4904]: I0214 12:15:01.610683 4904 generic.go:334] "Generic (PLEG): container finished" podID="e3558d0e-d510-4267-8937-f4d3cfe5d178" containerID="d32ebaa84cc32a704d4a48a347e292891639755755ca209d626d6c5a92ffa51f" exitCode=0 Feb 14 12:15:01 crc kubenswrapper[4904]: I0214 12:15:01.610794 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517855-lsnzv" event={"ID":"e3558d0e-d510-4267-8937-f4d3cfe5d178","Type":"ContainerDied","Data":"d32ebaa84cc32a704d4a48a347e292891639755755ca209d626d6c5a92ffa51f"} Feb 14 12:15:01 crc kubenswrapper[4904]: I0214 12:15:01.611245 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517855-lsnzv" event={"ID":"e3558d0e-d510-4267-8937-f4d3cfe5d178","Type":"ContainerStarted","Data":"b103bb44f2278bc1bdbd87f1ae70c9257df2e033619bfc96ffeef2bb65dade84"} Feb 14 12:15:01 crc kubenswrapper[4904]: I0214 12:15:01.837233 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:15:01 crc kubenswrapper[4904]: E0214 12:15:01.837488 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:15:02 crc kubenswrapper[4904]: I0214 12:15:02.971367 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517855-lsnzv" Feb 14 12:15:03 crc kubenswrapper[4904]: I0214 12:15:03.095894 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e3558d0e-d510-4267-8937-f4d3cfe5d178-config-volume\") pod \"e3558d0e-d510-4267-8937-f4d3cfe5d178\" (UID: \"e3558d0e-d510-4267-8937-f4d3cfe5d178\") " Feb 14 12:15:03 crc kubenswrapper[4904]: I0214 12:15:03.095949 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckjjv\" (UniqueName: \"kubernetes.io/projected/e3558d0e-d510-4267-8937-f4d3cfe5d178-kube-api-access-ckjjv\") pod \"e3558d0e-d510-4267-8937-f4d3cfe5d178\" (UID: \"e3558d0e-d510-4267-8937-f4d3cfe5d178\") " Feb 14 12:15:03 crc kubenswrapper[4904]: I0214 12:15:03.096002 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e3558d0e-d510-4267-8937-f4d3cfe5d178-secret-volume\") pod \"e3558d0e-d510-4267-8937-f4d3cfe5d178\" (UID: \"e3558d0e-d510-4267-8937-f4d3cfe5d178\") " Feb 14 12:15:03 crc kubenswrapper[4904]: I0214 12:15:03.096748 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3558d0e-d510-4267-8937-f4d3cfe5d178-config-volume" (OuterVolumeSpecName: "config-volume") pod "e3558d0e-d510-4267-8937-f4d3cfe5d178" (UID: "e3558d0e-d510-4267-8937-f4d3cfe5d178"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 12:15:03 crc kubenswrapper[4904]: I0214 12:15:03.102055 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3558d0e-d510-4267-8937-f4d3cfe5d178-kube-api-access-ckjjv" (OuterVolumeSpecName: "kube-api-access-ckjjv") pod "e3558d0e-d510-4267-8937-f4d3cfe5d178" (UID: "e3558d0e-d510-4267-8937-f4d3cfe5d178"). InnerVolumeSpecName "kube-api-access-ckjjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:15:03 crc kubenswrapper[4904]: I0214 12:15:03.103587 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3558d0e-d510-4267-8937-f4d3cfe5d178-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e3558d0e-d510-4267-8937-f4d3cfe5d178" (UID: "e3558d0e-d510-4267-8937-f4d3cfe5d178"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 12:15:03 crc kubenswrapper[4904]: I0214 12:15:03.197745 4904 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e3558d0e-d510-4267-8937-f4d3cfe5d178-config-volume\") on node \"crc\" DevicePath \"\"" Feb 14 12:15:03 crc kubenswrapper[4904]: I0214 12:15:03.197779 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckjjv\" (UniqueName: \"kubernetes.io/projected/e3558d0e-d510-4267-8937-f4d3cfe5d178-kube-api-access-ckjjv\") on node \"crc\" DevicePath \"\"" Feb 14 12:15:03 crc kubenswrapper[4904]: I0214 12:15:03.197794 4904 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e3558d0e-d510-4267-8937-f4d3cfe5d178-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 14 12:15:03 crc kubenswrapper[4904]: I0214 12:15:03.628071 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517855-lsnzv" event={"ID":"e3558d0e-d510-4267-8937-f4d3cfe5d178","Type":"ContainerDied","Data":"b103bb44f2278bc1bdbd87f1ae70c9257df2e033619bfc96ffeef2bb65dade84"} Feb 14 12:15:03 crc kubenswrapper[4904]: I0214 12:15:03.628110 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b103bb44f2278bc1bdbd87f1ae70c9257df2e033619bfc96ffeef2bb65dade84" Feb 14 12:15:03 crc kubenswrapper[4904]: I0214 12:15:03.628158 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517855-lsnzv" Feb 14 12:15:04 crc kubenswrapper[4904]: I0214 12:15:04.046362 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl"] Feb 14 12:15:04 crc kubenswrapper[4904]: I0214 12:15:04.053096 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517810-4d9dl"] Feb 14 12:15:05 crc kubenswrapper[4904]: I0214 12:15:05.848038 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71c31cb0-4057-4263-96d8-8938381f2b5b" path="/var/lib/kubelet/pods/71c31cb0-4057-4263-96d8-8938381f2b5b/volumes" Feb 14 12:15:12 crc kubenswrapper[4904]: I0214 12:15:12.836151 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:15:12 crc kubenswrapper[4904]: E0214 12:15:12.836944 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:15:16 crc kubenswrapper[4904]: I0214 12:15:16.616421 4904 scope.go:117] "RemoveContainer" containerID="b0ce3b7f36a57048b805388454f435ef2c7144169b88c9384c3de7802534c066" Feb 14 12:15:25 crc kubenswrapper[4904]: I0214 12:15:25.837511 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:15:26 crc kubenswrapper[4904]: I0214 12:15:26.810527 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"3d2a6006330ed7d1f70b590dd1d419d144dc78eacbc6aa350eb18ce68ec3b747"} Feb 14 12:17:46 crc kubenswrapper[4904]: I0214 12:17:46.383146 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:17:46 crc kubenswrapper[4904]: I0214 12:17:46.383594 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:18:16 crc kubenswrapper[4904]: I0214 12:18:16.383149 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:18:16 crc kubenswrapper[4904]: I0214 12:18:16.383948 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:18:17 crc kubenswrapper[4904]: I0214 12:18:17.175154 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hb7fd"] Feb 14 12:18:17 crc kubenswrapper[4904]: E0214 12:18:17.175602 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3558d0e-d510-4267-8937-f4d3cfe5d178" containerName="collect-profiles" Feb 14 12:18:17 crc kubenswrapper[4904]: I0214 12:18:17.175623 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3558d0e-d510-4267-8937-f4d3cfe5d178" containerName="collect-profiles" Feb 14 12:18:17 crc kubenswrapper[4904]: I0214 12:18:17.176000 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3558d0e-d510-4267-8937-f4d3cfe5d178" containerName="collect-profiles" Feb 14 12:18:17 crc kubenswrapper[4904]: I0214 12:18:17.177603 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hb7fd" Feb 14 12:18:17 crc kubenswrapper[4904]: I0214 12:18:17.188774 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hb7fd"] Feb 14 12:18:17 crc kubenswrapper[4904]: I0214 12:18:17.297126 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpc8k\" (UniqueName: \"kubernetes.io/projected/7abcd702-2bed-483a-8971-d665f658982d-kube-api-access-tpc8k\") pod \"redhat-operators-hb7fd\" (UID: \"7abcd702-2bed-483a-8971-d665f658982d\") " pod="openshift-marketplace/redhat-operators-hb7fd" Feb 14 12:18:17 crc kubenswrapper[4904]: I0214 12:18:17.297180 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7abcd702-2bed-483a-8971-d665f658982d-utilities\") pod \"redhat-operators-hb7fd\" (UID: \"7abcd702-2bed-483a-8971-d665f658982d\") " pod="openshift-marketplace/redhat-operators-hb7fd" Feb 14 12:18:17 crc kubenswrapper[4904]: I0214 12:18:17.297313 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7abcd702-2bed-483a-8971-d665f658982d-catalog-content\") pod \"redhat-operators-hb7fd\" (UID: \"7abcd702-2bed-483a-8971-d665f658982d\") " pod="openshift-marketplace/redhat-operators-hb7fd" Feb 14 12:18:17 crc kubenswrapper[4904]: I0214 12:18:17.398949 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7abcd702-2bed-483a-8971-d665f658982d-catalog-content\") pod \"redhat-operators-hb7fd\" (UID: \"7abcd702-2bed-483a-8971-d665f658982d\") " pod="openshift-marketplace/redhat-operators-hb7fd" Feb 14 12:18:17 crc kubenswrapper[4904]: I0214 12:18:17.399084 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpc8k\" (UniqueName: \"kubernetes.io/projected/7abcd702-2bed-483a-8971-d665f658982d-kube-api-access-tpc8k\") pod \"redhat-operators-hb7fd\" (UID: \"7abcd702-2bed-483a-8971-d665f658982d\") " pod="openshift-marketplace/redhat-operators-hb7fd" Feb 14 12:18:17 crc kubenswrapper[4904]: I0214 12:18:17.399107 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7abcd702-2bed-483a-8971-d665f658982d-utilities\") pod \"redhat-operators-hb7fd\" (UID: \"7abcd702-2bed-483a-8971-d665f658982d\") " pod="openshift-marketplace/redhat-operators-hb7fd" Feb 14 12:18:17 crc kubenswrapper[4904]: I0214 12:18:17.399536 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7abcd702-2bed-483a-8971-d665f658982d-utilities\") pod \"redhat-operators-hb7fd\" (UID: \"7abcd702-2bed-483a-8971-d665f658982d\") " pod="openshift-marketplace/redhat-operators-hb7fd" Feb 14 12:18:17 crc kubenswrapper[4904]: I0214 12:18:17.399937 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7abcd702-2bed-483a-8971-d665f658982d-catalog-content\") pod \"redhat-operators-hb7fd\" (UID: \"7abcd702-2bed-483a-8971-d665f658982d\") " pod="openshift-marketplace/redhat-operators-hb7fd" Feb 14 12:18:17 crc kubenswrapper[4904]: I0214 12:18:17.427782 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpc8k\" (UniqueName: \"kubernetes.io/projected/7abcd702-2bed-483a-8971-d665f658982d-kube-api-access-tpc8k\") pod \"redhat-operators-hb7fd\" (UID: \"7abcd702-2bed-483a-8971-d665f658982d\") " pod="openshift-marketplace/redhat-operators-hb7fd" Feb 14 12:18:17 crc kubenswrapper[4904]: I0214 12:18:17.499700 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hb7fd" Feb 14 12:18:17 crc kubenswrapper[4904]: I0214 12:18:17.956076 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hb7fd"] Feb 14 12:18:18 crc kubenswrapper[4904]: I0214 12:18:18.353114 4904 generic.go:334] "Generic (PLEG): container finished" podID="7abcd702-2bed-483a-8971-d665f658982d" containerID="9b22f417d84d8fb6e1bbf1caf7dec0a52fdd3d58d8bcf6fd8675fc653a3927f5" exitCode=0 Feb 14 12:18:18 crc kubenswrapper[4904]: I0214 12:18:18.353152 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hb7fd" event={"ID":"7abcd702-2bed-483a-8971-d665f658982d","Type":"ContainerDied","Data":"9b22f417d84d8fb6e1bbf1caf7dec0a52fdd3d58d8bcf6fd8675fc653a3927f5"} Feb 14 12:18:18 crc kubenswrapper[4904]: I0214 12:18:18.353174 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hb7fd" event={"ID":"7abcd702-2bed-483a-8971-d665f658982d","Type":"ContainerStarted","Data":"2ddf50445669913a06d391e69208180b293781266b48246b5a3039679a078eb6"} Feb 14 12:18:19 crc kubenswrapper[4904]: I0214 12:18:19.361483 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hb7fd" event={"ID":"7abcd702-2bed-483a-8971-d665f658982d","Type":"ContainerStarted","Data":"90981272883fdfd563981e01be32492459498ef57234f68e1ff72a68e4b0e3f1"} Feb 14 12:18:24 crc kubenswrapper[4904]: I0214 12:18:24.405409 4904 generic.go:334] "Generic (PLEG): container finished" podID="7abcd702-2bed-483a-8971-d665f658982d" containerID="90981272883fdfd563981e01be32492459498ef57234f68e1ff72a68e4b0e3f1" exitCode=0 Feb 14 12:18:24 crc kubenswrapper[4904]: I0214 12:18:24.405482 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hb7fd" event={"ID":"7abcd702-2bed-483a-8971-d665f658982d","Type":"ContainerDied","Data":"90981272883fdfd563981e01be32492459498ef57234f68e1ff72a68e4b0e3f1"} Feb 14 12:18:24 crc kubenswrapper[4904]: I0214 12:18:24.410540 4904 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 14 12:18:25 crc kubenswrapper[4904]: I0214 12:18:25.414702 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hb7fd" event={"ID":"7abcd702-2bed-483a-8971-d665f658982d","Type":"ContainerStarted","Data":"35f2bb539a66794e8aaca778e89d26d710c73ff9944442eab7a6cc522928d984"} Feb 14 12:18:25 crc kubenswrapper[4904]: I0214 12:18:25.434058 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hb7fd" podStartSLOduration=2.003382551 podStartE2EDuration="8.43403566s" podCreationTimestamp="2026-02-14 12:18:17 +0000 UTC" firstStartedPulling="2026-02-14 12:18:18.354491587 +0000 UTC m=+4089.167256248" lastFinishedPulling="2026-02-14 12:18:24.785144686 +0000 UTC m=+4095.597909357" observedRunningTime="2026-02-14 12:18:25.429966619 +0000 UTC m=+4096.242731280" watchObservedRunningTime="2026-02-14 12:18:25.43403566 +0000 UTC m=+4096.246800321" Feb 14 12:18:27 crc kubenswrapper[4904]: I0214 12:18:27.501193 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hb7fd" Feb 14 12:18:27 crc kubenswrapper[4904]: I0214 12:18:27.501526 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hb7fd" Feb 14 12:18:28 crc kubenswrapper[4904]: I0214 12:18:28.545553 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hb7fd" podUID="7abcd702-2bed-483a-8971-d665f658982d" containerName="registry-server" probeResult="failure" output=< Feb 14 12:18:28 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 12:18:28 crc kubenswrapper[4904]: > Feb 14 12:18:37 crc kubenswrapper[4904]: I0214 12:18:37.551176 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hb7fd" Feb 14 12:18:37 crc kubenswrapper[4904]: I0214 12:18:37.598210 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hb7fd" Feb 14 12:18:37 crc kubenswrapper[4904]: I0214 12:18:37.784615 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hb7fd"] Feb 14 12:18:39 crc kubenswrapper[4904]: I0214 12:18:39.528581 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hb7fd" podUID="7abcd702-2bed-483a-8971-d665f658982d" containerName="registry-server" containerID="cri-o://35f2bb539a66794e8aaca778e89d26d710c73ff9944442eab7a6cc522928d984" gracePeriod=2 Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.418453 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hb7fd" Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.533878 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7abcd702-2bed-483a-8971-d665f658982d-utilities\") pod \"7abcd702-2bed-483a-8971-d665f658982d\" (UID: \"7abcd702-2bed-483a-8971-d665f658982d\") " Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.534043 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7abcd702-2bed-483a-8971-d665f658982d-catalog-content\") pod \"7abcd702-2bed-483a-8971-d665f658982d\" (UID: \"7abcd702-2bed-483a-8971-d665f658982d\") " Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.534133 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpc8k\" (UniqueName: \"kubernetes.io/projected/7abcd702-2bed-483a-8971-d665f658982d-kube-api-access-tpc8k\") pod \"7abcd702-2bed-483a-8971-d665f658982d\" (UID: \"7abcd702-2bed-483a-8971-d665f658982d\") " Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.535602 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7abcd702-2bed-483a-8971-d665f658982d-utilities" (OuterVolumeSpecName: "utilities") pod "7abcd702-2bed-483a-8971-d665f658982d" (UID: "7abcd702-2bed-483a-8971-d665f658982d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.539896 4904 generic.go:334] "Generic (PLEG): container finished" podID="7abcd702-2bed-483a-8971-d665f658982d" containerID="35f2bb539a66794e8aaca778e89d26d710c73ff9944442eab7a6cc522928d984" exitCode=0 Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.539939 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hb7fd" event={"ID":"7abcd702-2bed-483a-8971-d665f658982d","Type":"ContainerDied","Data":"35f2bb539a66794e8aaca778e89d26d710c73ff9944442eab7a6cc522928d984"} Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.539965 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hb7fd" event={"ID":"7abcd702-2bed-483a-8971-d665f658982d","Type":"ContainerDied","Data":"2ddf50445669913a06d391e69208180b293781266b48246b5a3039679a078eb6"} Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.539980 4904 scope.go:117] "RemoveContainer" containerID="35f2bb539a66794e8aaca778e89d26d710c73ff9944442eab7a6cc522928d984" Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.540114 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hb7fd" Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.541590 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7abcd702-2bed-483a-8971-d665f658982d-kube-api-access-tpc8k" (OuterVolumeSpecName: "kube-api-access-tpc8k") pod "7abcd702-2bed-483a-8971-d665f658982d" (UID: "7abcd702-2bed-483a-8971-d665f658982d"). InnerVolumeSpecName "kube-api-access-tpc8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.595487 4904 scope.go:117] "RemoveContainer" containerID="90981272883fdfd563981e01be32492459498ef57234f68e1ff72a68e4b0e3f1" Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.616448 4904 scope.go:117] "RemoveContainer" containerID="9b22f417d84d8fb6e1bbf1caf7dec0a52fdd3d58d8bcf6fd8675fc653a3927f5" Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.636343 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpc8k\" (UniqueName: \"kubernetes.io/projected/7abcd702-2bed-483a-8971-d665f658982d-kube-api-access-tpc8k\") on node \"crc\" DevicePath \"\"" Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.636372 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7abcd702-2bed-483a-8971-d665f658982d-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.663219 4904 scope.go:117] "RemoveContainer" containerID="35f2bb539a66794e8aaca778e89d26d710c73ff9944442eab7a6cc522928d984" Feb 14 12:18:40 crc kubenswrapper[4904]: E0214 12:18:40.663701 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35f2bb539a66794e8aaca778e89d26d710c73ff9944442eab7a6cc522928d984\": container with ID starting with 35f2bb539a66794e8aaca778e89d26d710c73ff9944442eab7a6cc522928d984 not found: ID does not exist" containerID="35f2bb539a66794e8aaca778e89d26d710c73ff9944442eab7a6cc522928d984" Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.663730 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35f2bb539a66794e8aaca778e89d26d710c73ff9944442eab7a6cc522928d984"} err="failed to get container status \"35f2bb539a66794e8aaca778e89d26d710c73ff9944442eab7a6cc522928d984\": rpc error: code = NotFound desc = could not find container \"35f2bb539a66794e8aaca778e89d26d710c73ff9944442eab7a6cc522928d984\": container with ID starting with 35f2bb539a66794e8aaca778e89d26d710c73ff9944442eab7a6cc522928d984 not found: ID does not exist" Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.663750 4904 scope.go:117] "RemoveContainer" containerID="90981272883fdfd563981e01be32492459498ef57234f68e1ff72a68e4b0e3f1" Feb 14 12:18:40 crc kubenswrapper[4904]: E0214 12:18:40.664191 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90981272883fdfd563981e01be32492459498ef57234f68e1ff72a68e4b0e3f1\": container with ID starting with 90981272883fdfd563981e01be32492459498ef57234f68e1ff72a68e4b0e3f1 not found: ID does not exist" containerID="90981272883fdfd563981e01be32492459498ef57234f68e1ff72a68e4b0e3f1" Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.664211 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90981272883fdfd563981e01be32492459498ef57234f68e1ff72a68e4b0e3f1"} err="failed to get container status \"90981272883fdfd563981e01be32492459498ef57234f68e1ff72a68e4b0e3f1\": rpc error: code = NotFound desc = could not find container \"90981272883fdfd563981e01be32492459498ef57234f68e1ff72a68e4b0e3f1\": container with ID starting with 90981272883fdfd563981e01be32492459498ef57234f68e1ff72a68e4b0e3f1 not found: ID does not exist" Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.664233 4904 scope.go:117] "RemoveContainer" containerID="9b22f417d84d8fb6e1bbf1caf7dec0a52fdd3d58d8bcf6fd8675fc653a3927f5" Feb 14 12:18:40 crc kubenswrapper[4904]: E0214 12:18:40.664659 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b22f417d84d8fb6e1bbf1caf7dec0a52fdd3d58d8bcf6fd8675fc653a3927f5\": container with ID starting with 9b22f417d84d8fb6e1bbf1caf7dec0a52fdd3d58d8bcf6fd8675fc653a3927f5 not found: ID does not exist" containerID="9b22f417d84d8fb6e1bbf1caf7dec0a52fdd3d58d8bcf6fd8675fc653a3927f5" Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.664686 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b22f417d84d8fb6e1bbf1caf7dec0a52fdd3d58d8bcf6fd8675fc653a3927f5"} err="failed to get container status \"9b22f417d84d8fb6e1bbf1caf7dec0a52fdd3d58d8bcf6fd8675fc653a3927f5\": rpc error: code = NotFound desc = could not find container \"9b22f417d84d8fb6e1bbf1caf7dec0a52fdd3d58d8bcf6fd8675fc653a3927f5\": container with ID starting with 9b22f417d84d8fb6e1bbf1caf7dec0a52fdd3d58d8bcf6fd8675fc653a3927f5 not found: ID does not exist" Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.668104 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7abcd702-2bed-483a-8971-d665f658982d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7abcd702-2bed-483a-8971-d665f658982d" (UID: "7abcd702-2bed-483a-8971-d665f658982d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.738183 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7abcd702-2bed-483a-8971-d665f658982d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.878030 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hb7fd"] Feb 14 12:18:40 crc kubenswrapper[4904]: I0214 12:18:40.884935 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hb7fd"] Feb 14 12:18:41 crc kubenswrapper[4904]: I0214 12:18:41.861323 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7abcd702-2bed-483a-8971-d665f658982d" path="/var/lib/kubelet/pods/7abcd702-2bed-483a-8971-d665f658982d/volumes" Feb 14 12:18:46 crc kubenswrapper[4904]: I0214 12:18:46.382601 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:18:46 crc kubenswrapper[4904]: I0214 12:18:46.383276 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:18:46 crc kubenswrapper[4904]: I0214 12:18:46.383336 4904 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 12:18:46 crc kubenswrapper[4904]: I0214 12:18:46.384322 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3d2a6006330ed7d1f70b590dd1d419d144dc78eacbc6aa350eb18ce68ec3b747"} pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 14 12:18:46 crc kubenswrapper[4904]: I0214 12:18:46.384384 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" containerID="cri-o://3d2a6006330ed7d1f70b590dd1d419d144dc78eacbc6aa350eb18ce68ec3b747" gracePeriod=600 Feb 14 12:18:46 crc kubenswrapper[4904]: I0214 12:18:46.586380 4904 generic.go:334] "Generic (PLEG): container finished" podID="7790d276-2506-4641-8c22-67e6441417f0" containerID="3d2a6006330ed7d1f70b590dd1d419d144dc78eacbc6aa350eb18ce68ec3b747" exitCode=0 Feb 14 12:18:46 crc kubenswrapper[4904]: I0214 12:18:46.586905 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerDied","Data":"3d2a6006330ed7d1f70b590dd1d419d144dc78eacbc6aa350eb18ce68ec3b747"} Feb 14 12:18:46 crc kubenswrapper[4904]: I0214 12:18:46.586940 4904 scope.go:117] "RemoveContainer" containerID="38467b480e0c48b7cc109c85ecb3a383743659824e58f8e807631a4d85987664" Feb 14 12:18:46 crc kubenswrapper[4904]: E0214 12:18:46.594026 4904 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7790d276_2506_4641_8c22_67e6441417f0.slice/crio-3d2a6006330ed7d1f70b590dd1d419d144dc78eacbc6aa350eb18ce68ec3b747.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7790d276_2506_4641_8c22_67e6441417f0.slice/crio-conmon-3d2a6006330ed7d1f70b590dd1d419d144dc78eacbc6aa350eb18ce68ec3b747.scope\": RecentStats: unable to find data in memory cache]" Feb 14 12:18:47 crc kubenswrapper[4904]: I0214 12:18:47.596211 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28"} Feb 14 12:20:46 crc kubenswrapper[4904]: I0214 12:20:46.383193 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:20:46 crc kubenswrapper[4904]: I0214 12:20:46.384635 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:21:16 crc kubenswrapper[4904]: I0214 12:21:16.383044 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:21:16 crc kubenswrapper[4904]: I0214 12:21:16.383718 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:21:46 crc kubenswrapper[4904]: I0214 12:21:46.382767 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:21:46 crc kubenswrapper[4904]: I0214 12:21:46.383386 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:21:46 crc kubenswrapper[4904]: I0214 12:21:46.383436 4904 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 12:21:46 crc kubenswrapper[4904]: I0214 12:21:46.384222 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28"} pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 14 12:21:46 crc kubenswrapper[4904]: I0214 12:21:46.384276 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" containerID="cri-o://fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" gracePeriod=600 Feb 14 12:21:46 crc kubenswrapper[4904]: E0214 12:21:46.503650 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:21:47 crc kubenswrapper[4904]: I0214 12:21:47.431575 4904 generic.go:334] "Generic (PLEG): container finished" podID="7790d276-2506-4641-8c22-67e6441417f0" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" exitCode=0 Feb 14 12:21:47 crc kubenswrapper[4904]: I0214 12:21:47.431642 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerDied","Data":"fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28"} Feb 14 12:21:47 crc kubenswrapper[4904]: I0214 12:21:47.431902 4904 scope.go:117] "RemoveContainer" containerID="3d2a6006330ed7d1f70b590dd1d419d144dc78eacbc6aa350eb18ce68ec3b747" Feb 14 12:21:47 crc kubenswrapper[4904]: I0214 12:21:47.432515 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:21:47 crc kubenswrapper[4904]: E0214 12:21:47.432754 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:22:02 crc kubenswrapper[4904]: I0214 12:22:02.836356 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:22:02 crc kubenswrapper[4904]: E0214 12:22:02.837239 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:22:15 crc kubenswrapper[4904]: I0214 12:22:15.836753 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:22:15 crc kubenswrapper[4904]: E0214 12:22:15.837584 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:22:29 crc kubenswrapper[4904]: I0214 12:22:29.842351 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:22:29 crc kubenswrapper[4904]: E0214 12:22:29.843147 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:22:43 crc kubenswrapper[4904]: I0214 12:22:43.836971 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:22:43 crc kubenswrapper[4904]: E0214 12:22:43.839177 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:22:57 crc kubenswrapper[4904]: I0214 12:22:57.837468 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:22:57 crc kubenswrapper[4904]: E0214 12:22:57.838365 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:23:09 crc kubenswrapper[4904]: I0214 12:23:09.845111 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:23:09 crc kubenswrapper[4904]: E0214 12:23:09.846259 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:23:23 crc kubenswrapper[4904]: I0214 12:23:23.836576 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:23:23 crc kubenswrapper[4904]: E0214 12:23:23.837343 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:23:36 crc kubenswrapper[4904]: I0214 12:23:36.836434 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:23:36 crc kubenswrapper[4904]: E0214 12:23:36.837150 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:23:48 crc kubenswrapper[4904]: I0214 12:23:48.836512 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:23:48 crc kubenswrapper[4904]: E0214 12:23:48.837149 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:23:51 crc kubenswrapper[4904]: I0214 12:23:51.173652 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jnt5g"] Feb 14 12:23:51 crc kubenswrapper[4904]: E0214 12:23:51.174562 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7abcd702-2bed-483a-8971-d665f658982d" containerName="registry-server" Feb 14 12:23:51 crc kubenswrapper[4904]: I0214 12:23:51.174576 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="7abcd702-2bed-483a-8971-d665f658982d" containerName="registry-server" Feb 14 12:23:51 crc kubenswrapper[4904]: E0214 12:23:51.174598 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7abcd702-2bed-483a-8971-d665f658982d" containerName="extract-content" Feb 14 12:23:51 crc kubenswrapper[4904]: I0214 12:23:51.174604 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="7abcd702-2bed-483a-8971-d665f658982d" containerName="extract-content" Feb 14 12:23:51 crc kubenswrapper[4904]: E0214 12:23:51.174611 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7abcd702-2bed-483a-8971-d665f658982d" containerName="extract-utilities" Feb 14 12:23:51 crc kubenswrapper[4904]: I0214 12:23:51.174617 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="7abcd702-2bed-483a-8971-d665f658982d" containerName="extract-utilities" Feb 14 12:23:51 crc kubenswrapper[4904]: I0214 12:23:51.174853 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="7abcd702-2bed-483a-8971-d665f658982d" containerName="registry-server" Feb 14 12:23:51 crc kubenswrapper[4904]: I0214 12:23:51.176136 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jnt5g" Feb 14 12:23:51 crc kubenswrapper[4904]: I0214 12:23:51.190586 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jnt5g"] Feb 14 12:23:51 crc kubenswrapper[4904]: I0214 12:23:51.217856 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c98c78a6-e8c6-499e-bfbb-9bb9032da5a2-utilities\") pod \"community-operators-jnt5g\" (UID: \"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2\") " pod="openshift-marketplace/community-operators-jnt5g" Feb 14 12:23:51 crc kubenswrapper[4904]: I0214 12:23:51.218053 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pml8c\" (UniqueName: \"kubernetes.io/projected/c98c78a6-e8c6-499e-bfbb-9bb9032da5a2-kube-api-access-pml8c\") pod \"community-operators-jnt5g\" (UID: \"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2\") " pod="openshift-marketplace/community-operators-jnt5g" Feb 14 12:23:51 crc kubenswrapper[4904]: I0214 12:23:51.218095 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c98c78a6-e8c6-499e-bfbb-9bb9032da5a2-catalog-content\") pod \"community-operators-jnt5g\" (UID: \"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2\") " pod="openshift-marketplace/community-operators-jnt5g" Feb 14 12:23:51 crc kubenswrapper[4904]: I0214 12:23:51.320367 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c98c78a6-e8c6-499e-bfbb-9bb9032da5a2-utilities\") pod \"community-operators-jnt5g\" (UID: \"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2\") " pod="openshift-marketplace/community-operators-jnt5g" Feb 14 12:23:51 crc kubenswrapper[4904]: I0214 12:23:51.320542 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pml8c\" (UniqueName: \"kubernetes.io/projected/c98c78a6-e8c6-499e-bfbb-9bb9032da5a2-kube-api-access-pml8c\") pod \"community-operators-jnt5g\" (UID: \"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2\") " pod="openshift-marketplace/community-operators-jnt5g" Feb 14 12:23:51 crc kubenswrapper[4904]: I0214 12:23:51.320582 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c98c78a6-e8c6-499e-bfbb-9bb9032da5a2-catalog-content\") pod \"community-operators-jnt5g\" (UID: \"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2\") " pod="openshift-marketplace/community-operators-jnt5g" Feb 14 12:23:51 crc kubenswrapper[4904]: I0214 12:23:51.321078 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c98c78a6-e8c6-499e-bfbb-9bb9032da5a2-utilities\") pod \"community-operators-jnt5g\" (UID: \"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2\") " pod="openshift-marketplace/community-operators-jnt5g" Feb 14 12:23:51 crc kubenswrapper[4904]: I0214 12:23:51.321228 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c98c78a6-e8c6-499e-bfbb-9bb9032da5a2-catalog-content\") pod \"community-operators-jnt5g\" (UID: \"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2\") " pod="openshift-marketplace/community-operators-jnt5g" Feb 14 12:23:51 crc kubenswrapper[4904]: I0214 12:23:51.348715 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pml8c\" (UniqueName: \"kubernetes.io/projected/c98c78a6-e8c6-499e-bfbb-9bb9032da5a2-kube-api-access-pml8c\") pod \"community-operators-jnt5g\" (UID: \"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2\") " pod="openshift-marketplace/community-operators-jnt5g" Feb 14 12:23:51 crc kubenswrapper[4904]: I0214 12:23:51.494928 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jnt5g" Feb 14 12:23:52 crc kubenswrapper[4904]: I0214 12:23:52.098253 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jnt5g"] Feb 14 12:23:52 crc kubenswrapper[4904]: I0214 12:23:52.436922 4904 generic.go:334] "Generic (PLEG): container finished" podID="c98c78a6-e8c6-499e-bfbb-9bb9032da5a2" containerID="3606ae176fedd4ed3ffa437b0a190ed83fa4364e87ceac0d0f9aba4e63aa1544" exitCode=0 Feb 14 12:23:52 crc kubenswrapper[4904]: I0214 12:23:52.437013 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnt5g" event={"ID":"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2","Type":"ContainerDied","Data":"3606ae176fedd4ed3ffa437b0a190ed83fa4364e87ceac0d0f9aba4e63aa1544"} Feb 14 12:23:52 crc kubenswrapper[4904]: I0214 12:23:52.437261 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnt5g" event={"ID":"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2","Type":"ContainerStarted","Data":"eaf7b7fee4d10a489b78c35890eff5b949bbb449afebaa70d15cd5c21c1281fe"} Feb 14 12:23:52 crc kubenswrapper[4904]: I0214 12:23:52.438851 4904 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 14 12:23:54 crc kubenswrapper[4904]: I0214 12:23:54.465810 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnt5g" event={"ID":"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2","Type":"ContainerStarted","Data":"b8d431a88e225f3fc47990ea4a092fd7e988ad493bae6869919be33d29c1db14"} Feb 14 12:23:55 crc kubenswrapper[4904]: I0214 12:23:55.475962 4904 generic.go:334] "Generic (PLEG): container finished" podID="c98c78a6-e8c6-499e-bfbb-9bb9032da5a2" containerID="b8d431a88e225f3fc47990ea4a092fd7e988ad493bae6869919be33d29c1db14" exitCode=0 Feb 14 12:23:55 crc kubenswrapper[4904]: I0214 12:23:55.476009 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnt5g" event={"ID":"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2","Type":"ContainerDied","Data":"b8d431a88e225f3fc47990ea4a092fd7e988ad493bae6869919be33d29c1db14"} Feb 14 12:23:57 crc kubenswrapper[4904]: I0214 12:23:57.494866 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnt5g" event={"ID":"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2","Type":"ContainerStarted","Data":"5954664006b62e9791c364fdbf0027d3d0de0fd17ecd98bb2d9dfc2e8c870b45"} Feb 14 12:23:57 crc kubenswrapper[4904]: I0214 12:23:57.521041 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jnt5g" podStartSLOduration=3.117720893 podStartE2EDuration="6.521023534s" podCreationTimestamp="2026-02-14 12:23:51 +0000 UTC" firstStartedPulling="2026-02-14 12:23:52.438511716 +0000 UTC m=+4423.251276377" lastFinishedPulling="2026-02-14 12:23:55.841814357 +0000 UTC m=+4426.654579018" observedRunningTime="2026-02-14 12:23:57.520738747 +0000 UTC m=+4428.333503408" watchObservedRunningTime="2026-02-14 12:23:57.521023534 +0000 UTC m=+4428.333788195" Feb 14 12:24:01 crc kubenswrapper[4904]: I0214 12:24:01.496030 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jnt5g" Feb 14 12:24:01 crc kubenswrapper[4904]: I0214 12:24:01.496525 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jnt5g" Feb 14 12:24:01 crc kubenswrapper[4904]: I0214 12:24:01.541663 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jnt5g" Feb 14 12:24:01 crc kubenswrapper[4904]: I0214 12:24:01.603561 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jnt5g" Feb 14 12:24:01 crc kubenswrapper[4904]: I0214 12:24:01.783038 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jnt5g"] Feb 14 12:24:02 crc kubenswrapper[4904]: I0214 12:24:02.836292 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:24:02 crc kubenswrapper[4904]: E0214 12:24:02.836778 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:24:03 crc kubenswrapper[4904]: I0214 12:24:03.548354 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jnt5g" podUID="c98c78a6-e8c6-499e-bfbb-9bb9032da5a2" containerName="registry-server" containerID="cri-o://5954664006b62e9791c364fdbf0027d3d0de0fd17ecd98bb2d9dfc2e8c870b45" gracePeriod=2 Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.039373 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jnt5g" Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.094378 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c98c78a6-e8c6-499e-bfbb-9bb9032da5a2-catalog-content\") pod \"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2\" (UID: \"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2\") " Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.094505 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c98c78a6-e8c6-499e-bfbb-9bb9032da5a2-utilities\") pod \"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2\" (UID: \"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2\") " Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.094545 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pml8c\" (UniqueName: \"kubernetes.io/projected/c98c78a6-e8c6-499e-bfbb-9bb9032da5a2-kube-api-access-pml8c\") pod \"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2\" (UID: \"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2\") " Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.095354 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c98c78a6-e8c6-499e-bfbb-9bb9032da5a2-utilities" (OuterVolumeSpecName: "utilities") pod "c98c78a6-e8c6-499e-bfbb-9bb9032da5a2" (UID: "c98c78a6-e8c6-499e-bfbb-9bb9032da5a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.112722 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c98c78a6-e8c6-499e-bfbb-9bb9032da5a2-kube-api-access-pml8c" (OuterVolumeSpecName: "kube-api-access-pml8c") pod "c98c78a6-e8c6-499e-bfbb-9bb9032da5a2" (UID: "c98c78a6-e8c6-499e-bfbb-9bb9032da5a2"). InnerVolumeSpecName "kube-api-access-pml8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.145310 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c98c78a6-e8c6-499e-bfbb-9bb9032da5a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c98c78a6-e8c6-499e-bfbb-9bb9032da5a2" (UID: "c98c78a6-e8c6-499e-bfbb-9bb9032da5a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.196055 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c98c78a6-e8c6-499e-bfbb-9bb9032da5a2-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.196082 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c98c78a6-e8c6-499e-bfbb-9bb9032da5a2-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.196090 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pml8c\" (UniqueName: \"kubernetes.io/projected/c98c78a6-e8c6-499e-bfbb-9bb9032da5a2-kube-api-access-pml8c\") on node \"crc\" DevicePath \"\"" Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.557330 4904 generic.go:334] "Generic (PLEG): container finished" podID="c98c78a6-e8c6-499e-bfbb-9bb9032da5a2" containerID="5954664006b62e9791c364fdbf0027d3d0de0fd17ecd98bb2d9dfc2e8c870b45" exitCode=0 Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.557370 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnt5g" event={"ID":"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2","Type":"ContainerDied","Data":"5954664006b62e9791c364fdbf0027d3d0de0fd17ecd98bb2d9dfc2e8c870b45"} Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.557393 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnt5g" event={"ID":"c98c78a6-e8c6-499e-bfbb-9bb9032da5a2","Type":"ContainerDied","Data":"eaf7b7fee4d10a489b78c35890eff5b949bbb449afebaa70d15cd5c21c1281fe"} Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.557411 4904 scope.go:117] "RemoveContainer" containerID="5954664006b62e9791c364fdbf0027d3d0de0fd17ecd98bb2d9dfc2e8c870b45" Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.557956 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jnt5g" Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.576822 4904 scope.go:117] "RemoveContainer" containerID="b8d431a88e225f3fc47990ea4a092fd7e988ad493bae6869919be33d29c1db14" Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.598620 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jnt5g"] Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.614793 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jnt5g"] Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.630323 4904 scope.go:117] "RemoveContainer" containerID="3606ae176fedd4ed3ffa437b0a190ed83fa4364e87ceac0d0f9aba4e63aa1544" Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.656042 4904 scope.go:117] "RemoveContainer" containerID="5954664006b62e9791c364fdbf0027d3d0de0fd17ecd98bb2d9dfc2e8c870b45" Feb 14 12:24:04 crc kubenswrapper[4904]: E0214 12:24:04.660306 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5954664006b62e9791c364fdbf0027d3d0de0fd17ecd98bb2d9dfc2e8c870b45\": container with ID starting with 5954664006b62e9791c364fdbf0027d3d0de0fd17ecd98bb2d9dfc2e8c870b45 not found: ID does not exist" containerID="5954664006b62e9791c364fdbf0027d3d0de0fd17ecd98bb2d9dfc2e8c870b45" Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.660355 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5954664006b62e9791c364fdbf0027d3d0de0fd17ecd98bb2d9dfc2e8c870b45"} err="failed to get container status \"5954664006b62e9791c364fdbf0027d3d0de0fd17ecd98bb2d9dfc2e8c870b45\": rpc error: code = NotFound desc = could not find container \"5954664006b62e9791c364fdbf0027d3d0de0fd17ecd98bb2d9dfc2e8c870b45\": container with ID starting with 5954664006b62e9791c364fdbf0027d3d0de0fd17ecd98bb2d9dfc2e8c870b45 not found: ID does not exist" Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.660392 4904 scope.go:117] "RemoveContainer" containerID="b8d431a88e225f3fc47990ea4a092fd7e988ad493bae6869919be33d29c1db14" Feb 14 12:24:04 crc kubenswrapper[4904]: E0214 12:24:04.661618 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8d431a88e225f3fc47990ea4a092fd7e988ad493bae6869919be33d29c1db14\": container with ID starting with b8d431a88e225f3fc47990ea4a092fd7e988ad493bae6869919be33d29c1db14 not found: ID does not exist" containerID="b8d431a88e225f3fc47990ea4a092fd7e988ad493bae6869919be33d29c1db14" Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.661653 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8d431a88e225f3fc47990ea4a092fd7e988ad493bae6869919be33d29c1db14"} err="failed to get container status \"b8d431a88e225f3fc47990ea4a092fd7e988ad493bae6869919be33d29c1db14\": rpc error: code = NotFound desc = could not find container \"b8d431a88e225f3fc47990ea4a092fd7e988ad493bae6869919be33d29c1db14\": container with ID starting with b8d431a88e225f3fc47990ea4a092fd7e988ad493bae6869919be33d29c1db14 not found: ID does not exist" Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.661674 4904 scope.go:117] "RemoveContainer" containerID="3606ae176fedd4ed3ffa437b0a190ed83fa4364e87ceac0d0f9aba4e63aa1544" Feb 14 12:24:04 crc kubenswrapper[4904]: E0214 12:24:04.661954 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3606ae176fedd4ed3ffa437b0a190ed83fa4364e87ceac0d0f9aba4e63aa1544\": container with ID starting with 3606ae176fedd4ed3ffa437b0a190ed83fa4364e87ceac0d0f9aba4e63aa1544 not found: ID does not exist" containerID="3606ae176fedd4ed3ffa437b0a190ed83fa4364e87ceac0d0f9aba4e63aa1544" Feb 14 12:24:04 crc kubenswrapper[4904]: I0214 12:24:04.661985 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3606ae176fedd4ed3ffa437b0a190ed83fa4364e87ceac0d0f9aba4e63aa1544"} err="failed to get container status \"3606ae176fedd4ed3ffa437b0a190ed83fa4364e87ceac0d0f9aba4e63aa1544\": rpc error: code = NotFound desc = could not find container \"3606ae176fedd4ed3ffa437b0a190ed83fa4364e87ceac0d0f9aba4e63aa1544\": container with ID starting with 3606ae176fedd4ed3ffa437b0a190ed83fa4364e87ceac0d0f9aba4e63aa1544 not found: ID does not exist" Feb 14 12:24:05 crc kubenswrapper[4904]: I0214 12:24:05.848088 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c98c78a6-e8c6-499e-bfbb-9bb9032da5a2" path="/var/lib/kubelet/pods/c98c78a6-e8c6-499e-bfbb-9bb9032da5a2/volumes" Feb 14 12:24:15 crc kubenswrapper[4904]: I0214 12:24:15.836631 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:24:15 crc kubenswrapper[4904]: E0214 12:24:15.837315 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:24:28 crc kubenswrapper[4904]: I0214 12:24:28.840494 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:24:28 crc kubenswrapper[4904]: E0214 12:24:28.841273 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:24:42 crc kubenswrapper[4904]: I0214 12:24:42.836907 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:24:42 crc kubenswrapper[4904]: E0214 12:24:42.837632 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:24:54 crc kubenswrapper[4904]: I0214 12:24:54.836335 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:24:54 crc kubenswrapper[4904]: E0214 12:24:54.837110 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:24:55 crc kubenswrapper[4904]: I0214 12:24:55.369403 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-llmvg"] Feb 14 12:24:55 crc kubenswrapper[4904]: E0214 12:24:55.369879 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c98c78a6-e8c6-499e-bfbb-9bb9032da5a2" containerName="extract-utilities" Feb 14 12:24:55 crc kubenswrapper[4904]: I0214 12:24:55.369904 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="c98c78a6-e8c6-499e-bfbb-9bb9032da5a2" containerName="extract-utilities" Feb 14 12:24:55 crc kubenswrapper[4904]: E0214 12:24:55.369940 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c98c78a6-e8c6-499e-bfbb-9bb9032da5a2" containerName="extract-content" Feb 14 12:24:55 crc kubenswrapper[4904]: I0214 12:24:55.369949 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="c98c78a6-e8c6-499e-bfbb-9bb9032da5a2" containerName="extract-content" Feb 14 12:24:55 crc kubenswrapper[4904]: E0214 12:24:55.369974 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c98c78a6-e8c6-499e-bfbb-9bb9032da5a2" containerName="registry-server" Feb 14 12:24:55 crc kubenswrapper[4904]: I0214 12:24:55.369984 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="c98c78a6-e8c6-499e-bfbb-9bb9032da5a2" containerName="registry-server" Feb 14 12:24:55 crc kubenswrapper[4904]: I0214 12:24:55.370218 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="c98c78a6-e8c6-499e-bfbb-9bb9032da5a2" containerName="registry-server" Feb 14 12:24:55 crc kubenswrapper[4904]: I0214 12:24:55.372107 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-llmvg" Feb 14 12:24:55 crc kubenswrapper[4904]: I0214 12:24:55.390467 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-llmvg"] Feb 14 12:24:55 crc kubenswrapper[4904]: I0214 12:24:55.474726 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f278f66a-6395-4efb-bb4d-1628cd07d36c-utilities\") pod \"certified-operators-llmvg\" (UID: \"f278f66a-6395-4efb-bb4d-1628cd07d36c\") " pod="openshift-marketplace/certified-operators-llmvg" Feb 14 12:24:55 crc kubenswrapper[4904]: I0214 12:24:55.475191 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f278f66a-6395-4efb-bb4d-1628cd07d36c-catalog-content\") pod \"certified-operators-llmvg\" (UID: \"f278f66a-6395-4efb-bb4d-1628cd07d36c\") " pod="openshift-marketplace/certified-operators-llmvg" Feb 14 12:24:55 crc kubenswrapper[4904]: I0214 12:24:55.475329 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcbmk\" (UniqueName: \"kubernetes.io/projected/f278f66a-6395-4efb-bb4d-1628cd07d36c-kube-api-access-kcbmk\") pod \"certified-operators-llmvg\" (UID: \"f278f66a-6395-4efb-bb4d-1628cd07d36c\") " pod="openshift-marketplace/certified-operators-llmvg" Feb 14 12:24:55 crc kubenswrapper[4904]: I0214 12:24:55.577729 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f278f66a-6395-4efb-bb4d-1628cd07d36c-utilities\") pod \"certified-operators-llmvg\" (UID: \"f278f66a-6395-4efb-bb4d-1628cd07d36c\") " pod="openshift-marketplace/certified-operators-llmvg" Feb 14 12:24:55 crc kubenswrapper[4904]: I0214 12:24:55.577965 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f278f66a-6395-4efb-bb4d-1628cd07d36c-catalog-content\") pod \"certified-operators-llmvg\" (UID: \"f278f66a-6395-4efb-bb4d-1628cd07d36c\") " pod="openshift-marketplace/certified-operators-llmvg" Feb 14 12:24:55 crc kubenswrapper[4904]: I0214 12:24:55.577992 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcbmk\" (UniqueName: \"kubernetes.io/projected/f278f66a-6395-4efb-bb4d-1628cd07d36c-kube-api-access-kcbmk\") pod \"certified-operators-llmvg\" (UID: \"f278f66a-6395-4efb-bb4d-1628cd07d36c\") " pod="openshift-marketplace/certified-operators-llmvg" Feb 14 12:24:55 crc kubenswrapper[4904]: I0214 12:24:55.578425 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f278f66a-6395-4efb-bb4d-1628cd07d36c-utilities\") pod \"certified-operators-llmvg\" (UID: \"f278f66a-6395-4efb-bb4d-1628cd07d36c\") " pod="openshift-marketplace/certified-operators-llmvg" Feb 14 12:24:55 crc kubenswrapper[4904]: I0214 12:24:55.578900 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f278f66a-6395-4efb-bb4d-1628cd07d36c-catalog-content\") pod \"certified-operators-llmvg\" (UID: \"f278f66a-6395-4efb-bb4d-1628cd07d36c\") " pod="openshift-marketplace/certified-operators-llmvg" Feb 14 12:24:55 crc kubenswrapper[4904]: I0214 12:24:55.598250 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcbmk\" (UniqueName: \"kubernetes.io/projected/f278f66a-6395-4efb-bb4d-1628cd07d36c-kube-api-access-kcbmk\") pod \"certified-operators-llmvg\" (UID: \"f278f66a-6395-4efb-bb4d-1628cd07d36c\") " pod="openshift-marketplace/certified-operators-llmvg" Feb 14 12:24:55 crc kubenswrapper[4904]: I0214 12:24:55.708542 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-llmvg" Feb 14 12:24:56 crc kubenswrapper[4904]: I0214 12:24:56.225806 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-llmvg"] Feb 14 12:24:57 crc kubenswrapper[4904]: I0214 12:24:57.040483 4904 generic.go:334] "Generic (PLEG): container finished" podID="f278f66a-6395-4efb-bb4d-1628cd07d36c" containerID="d54ba8bc237f78705f8a731e2855bc12228c54d17c7003b3bc18909e9a37ba90" exitCode=0 Feb 14 12:24:57 crc kubenswrapper[4904]: I0214 12:24:57.040662 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llmvg" event={"ID":"f278f66a-6395-4efb-bb4d-1628cd07d36c","Type":"ContainerDied","Data":"d54ba8bc237f78705f8a731e2855bc12228c54d17c7003b3bc18909e9a37ba90"} Feb 14 12:24:57 crc kubenswrapper[4904]: I0214 12:24:57.040785 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llmvg" event={"ID":"f278f66a-6395-4efb-bb4d-1628cd07d36c","Type":"ContainerStarted","Data":"36d9f48834a83a9d2ba564d92e1bb98155796baf46eac2e28b8b5f9bcc7a1630"} Feb 14 12:24:58 crc kubenswrapper[4904]: I0214 12:24:58.050091 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llmvg" event={"ID":"f278f66a-6395-4efb-bb4d-1628cd07d36c","Type":"ContainerStarted","Data":"dfe73b638f86d4b4e444a6ff2bfa482fcaeb3e0fc9e8c545510d6815128c5143"} Feb 14 12:24:59 crc kubenswrapper[4904]: I0214 12:24:59.059342 4904 generic.go:334] "Generic (PLEG): container finished" podID="f278f66a-6395-4efb-bb4d-1628cd07d36c" containerID="dfe73b638f86d4b4e444a6ff2bfa482fcaeb3e0fc9e8c545510d6815128c5143" exitCode=0 Feb 14 12:24:59 crc kubenswrapper[4904]: I0214 12:24:59.059626 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llmvg" event={"ID":"f278f66a-6395-4efb-bb4d-1628cd07d36c","Type":"ContainerDied","Data":"dfe73b638f86d4b4e444a6ff2bfa482fcaeb3e0fc9e8c545510d6815128c5143"} Feb 14 12:25:00 crc kubenswrapper[4904]: I0214 12:25:00.071442 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llmvg" event={"ID":"f278f66a-6395-4efb-bb4d-1628cd07d36c","Type":"ContainerStarted","Data":"c142263705861455f3ed9f8cd3aede97f91cfb28d0a05a7746a0ca54351c4a85"} Feb 14 12:25:00 crc kubenswrapper[4904]: I0214 12:25:00.095262 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-llmvg" podStartSLOduration=2.688641451 podStartE2EDuration="5.095239395s" podCreationTimestamp="2026-02-14 12:24:55 +0000 UTC" firstStartedPulling="2026-02-14 12:24:57.042330449 +0000 UTC m=+4487.855095100" lastFinishedPulling="2026-02-14 12:24:59.448928383 +0000 UTC m=+4490.261693044" observedRunningTime="2026-02-14 12:25:00.087055702 +0000 UTC m=+4490.899820383" watchObservedRunningTime="2026-02-14 12:25:00.095239395 +0000 UTC m=+4490.908004056" Feb 14 12:25:05 crc kubenswrapper[4904]: I0214 12:25:05.708928 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-llmvg" Feb 14 12:25:05 crc kubenswrapper[4904]: I0214 12:25:05.709316 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-llmvg" Feb 14 12:25:05 crc kubenswrapper[4904]: I0214 12:25:05.772988 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-llmvg" Feb 14 12:25:06 crc kubenswrapper[4904]: I0214 12:25:06.188260 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-llmvg" Feb 14 12:25:06 crc kubenswrapper[4904]: I0214 12:25:06.237233 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-llmvg"] Feb 14 12:25:07 crc kubenswrapper[4904]: I0214 12:25:07.838691 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:25:07 crc kubenswrapper[4904]: E0214 12:25:07.840521 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:25:08 crc kubenswrapper[4904]: I0214 12:25:08.147509 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-llmvg" podUID="f278f66a-6395-4efb-bb4d-1628cd07d36c" containerName="registry-server" containerID="cri-o://c142263705861455f3ed9f8cd3aede97f91cfb28d0a05a7746a0ca54351c4a85" gracePeriod=2 Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.156062 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-llmvg" Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.157511 4904 generic.go:334] "Generic (PLEG): container finished" podID="f278f66a-6395-4efb-bb4d-1628cd07d36c" containerID="c142263705861455f3ed9f8cd3aede97f91cfb28d0a05a7746a0ca54351c4a85" exitCode=0 Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.157551 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llmvg" event={"ID":"f278f66a-6395-4efb-bb4d-1628cd07d36c","Type":"ContainerDied","Data":"c142263705861455f3ed9f8cd3aede97f91cfb28d0a05a7746a0ca54351c4a85"} Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.157576 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llmvg" event={"ID":"f278f66a-6395-4efb-bb4d-1628cd07d36c","Type":"ContainerDied","Data":"36d9f48834a83a9d2ba564d92e1bb98155796baf46eac2e28b8b5f9bcc7a1630"} Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.157606 4904 scope.go:117] "RemoveContainer" containerID="c142263705861455f3ed9f8cd3aede97f91cfb28d0a05a7746a0ca54351c4a85" Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.183652 4904 scope.go:117] "RemoveContainer" containerID="dfe73b638f86d4b4e444a6ff2bfa482fcaeb3e0fc9e8c545510d6815128c5143" Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.204386 4904 scope.go:117] "RemoveContainer" containerID="d54ba8bc237f78705f8a731e2855bc12228c54d17c7003b3bc18909e9a37ba90" Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.243349 4904 scope.go:117] "RemoveContainer" containerID="c142263705861455f3ed9f8cd3aede97f91cfb28d0a05a7746a0ca54351c4a85" Feb 14 12:25:09 crc kubenswrapper[4904]: E0214 12:25:09.243886 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c142263705861455f3ed9f8cd3aede97f91cfb28d0a05a7746a0ca54351c4a85\": container with ID starting with c142263705861455f3ed9f8cd3aede97f91cfb28d0a05a7746a0ca54351c4a85 not found: ID does not exist" containerID="c142263705861455f3ed9f8cd3aede97f91cfb28d0a05a7746a0ca54351c4a85" Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.243924 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c142263705861455f3ed9f8cd3aede97f91cfb28d0a05a7746a0ca54351c4a85"} err="failed to get container status \"c142263705861455f3ed9f8cd3aede97f91cfb28d0a05a7746a0ca54351c4a85\": rpc error: code = NotFound desc = could not find container \"c142263705861455f3ed9f8cd3aede97f91cfb28d0a05a7746a0ca54351c4a85\": container with ID starting with c142263705861455f3ed9f8cd3aede97f91cfb28d0a05a7746a0ca54351c4a85 not found: ID does not exist" Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.243950 4904 scope.go:117] "RemoveContainer" containerID="dfe73b638f86d4b4e444a6ff2bfa482fcaeb3e0fc9e8c545510d6815128c5143" Feb 14 12:25:09 crc kubenswrapper[4904]: E0214 12:25:09.244296 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfe73b638f86d4b4e444a6ff2bfa482fcaeb3e0fc9e8c545510d6815128c5143\": container with ID starting with dfe73b638f86d4b4e444a6ff2bfa482fcaeb3e0fc9e8c545510d6815128c5143 not found: ID does not exist" containerID="dfe73b638f86d4b4e444a6ff2bfa482fcaeb3e0fc9e8c545510d6815128c5143" Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.244340 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfe73b638f86d4b4e444a6ff2bfa482fcaeb3e0fc9e8c545510d6815128c5143"} err="failed to get container status \"dfe73b638f86d4b4e444a6ff2bfa482fcaeb3e0fc9e8c545510d6815128c5143\": rpc error: code = NotFound desc = could not find container \"dfe73b638f86d4b4e444a6ff2bfa482fcaeb3e0fc9e8c545510d6815128c5143\": container with ID starting with dfe73b638f86d4b4e444a6ff2bfa482fcaeb3e0fc9e8c545510d6815128c5143 not found: ID does not exist" Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.244363 4904 scope.go:117] "RemoveContainer" containerID="d54ba8bc237f78705f8a731e2855bc12228c54d17c7003b3bc18909e9a37ba90" Feb 14 12:25:09 crc kubenswrapper[4904]: E0214 12:25:09.244644 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d54ba8bc237f78705f8a731e2855bc12228c54d17c7003b3bc18909e9a37ba90\": container with ID starting with d54ba8bc237f78705f8a731e2855bc12228c54d17c7003b3bc18909e9a37ba90 not found: ID does not exist" containerID="d54ba8bc237f78705f8a731e2855bc12228c54d17c7003b3bc18909e9a37ba90" Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.244693 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d54ba8bc237f78705f8a731e2855bc12228c54d17c7003b3bc18909e9a37ba90"} err="failed to get container status \"d54ba8bc237f78705f8a731e2855bc12228c54d17c7003b3bc18909e9a37ba90\": rpc error: code = NotFound desc = could not find container \"d54ba8bc237f78705f8a731e2855bc12228c54d17c7003b3bc18909e9a37ba90\": container with ID starting with d54ba8bc237f78705f8a731e2855bc12228c54d17c7003b3bc18909e9a37ba90 not found: ID does not exist" Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.276148 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f278f66a-6395-4efb-bb4d-1628cd07d36c-utilities\") pod \"f278f66a-6395-4efb-bb4d-1628cd07d36c\" (UID: \"f278f66a-6395-4efb-bb4d-1628cd07d36c\") " Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.276287 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcbmk\" (UniqueName: \"kubernetes.io/projected/f278f66a-6395-4efb-bb4d-1628cd07d36c-kube-api-access-kcbmk\") pod \"f278f66a-6395-4efb-bb4d-1628cd07d36c\" (UID: \"f278f66a-6395-4efb-bb4d-1628cd07d36c\") " Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.276342 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f278f66a-6395-4efb-bb4d-1628cd07d36c-catalog-content\") pod \"f278f66a-6395-4efb-bb4d-1628cd07d36c\" (UID: \"f278f66a-6395-4efb-bb4d-1628cd07d36c\") " Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.277431 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f278f66a-6395-4efb-bb4d-1628cd07d36c-utilities" (OuterVolumeSpecName: "utilities") pod "f278f66a-6395-4efb-bb4d-1628cd07d36c" (UID: "f278f66a-6395-4efb-bb4d-1628cd07d36c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.282884 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f278f66a-6395-4efb-bb4d-1628cd07d36c-kube-api-access-kcbmk" (OuterVolumeSpecName: "kube-api-access-kcbmk") pod "f278f66a-6395-4efb-bb4d-1628cd07d36c" (UID: "f278f66a-6395-4efb-bb4d-1628cd07d36c"). InnerVolumeSpecName "kube-api-access-kcbmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.327616 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f278f66a-6395-4efb-bb4d-1628cd07d36c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f278f66a-6395-4efb-bb4d-1628cd07d36c" (UID: "f278f66a-6395-4efb-bb4d-1628cd07d36c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.378074 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f278f66a-6395-4efb-bb4d-1628cd07d36c-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.378110 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcbmk\" (UniqueName: \"kubernetes.io/projected/f278f66a-6395-4efb-bb4d-1628cd07d36c-kube-api-access-kcbmk\") on node \"crc\" DevicePath \"\"" Feb 14 12:25:09 crc kubenswrapper[4904]: I0214 12:25:09.378152 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f278f66a-6395-4efb-bb4d-1628cd07d36c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 12:25:10 crc kubenswrapper[4904]: I0214 12:25:10.167601 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-llmvg" Feb 14 12:25:10 crc kubenswrapper[4904]: I0214 12:25:10.199087 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-llmvg"] Feb 14 12:25:10 crc kubenswrapper[4904]: I0214 12:25:10.216727 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-llmvg"] Feb 14 12:25:11 crc kubenswrapper[4904]: I0214 12:25:11.855499 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f278f66a-6395-4efb-bb4d-1628cd07d36c" path="/var/lib/kubelet/pods/f278f66a-6395-4efb-bb4d-1628cd07d36c/volumes" Feb 14 12:25:15 crc kubenswrapper[4904]: E0214 12:25:15.581492 4904 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf278f66a_6395_4efb_bb4d_1628cd07d36c.slice\": RecentStats: unable to find data in memory cache]" Feb 14 12:25:22 crc kubenswrapper[4904]: I0214 12:25:22.836590 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:25:22 crc kubenswrapper[4904]: E0214 12:25:22.837363 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:25:25 crc kubenswrapper[4904]: E0214 12:25:25.819780 4904 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf278f66a_6395_4efb_bb4d_1628cd07d36c.slice\": RecentStats: unable to find data in memory cache]" Feb 14 12:25:36 crc kubenswrapper[4904]: E0214 12:25:36.030481 4904 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf278f66a_6395_4efb_bb4d_1628cd07d36c.slice\": RecentStats: unable to find data in memory cache]" Feb 14 12:25:37 crc kubenswrapper[4904]: I0214 12:25:37.836978 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:25:37 crc kubenswrapper[4904]: E0214 12:25:37.837949 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:25:46 crc kubenswrapper[4904]: E0214 12:25:46.246966 4904 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf278f66a_6395_4efb_bb4d_1628cd07d36c.slice\": RecentStats: unable to find data in memory cache]" Feb 14 12:25:52 crc kubenswrapper[4904]: I0214 12:25:52.838872 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:25:52 crc kubenswrapper[4904]: E0214 12:25:52.839638 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:25:56 crc kubenswrapper[4904]: E0214 12:25:56.484637 4904 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf278f66a_6395_4efb_bb4d_1628cd07d36c.slice\": RecentStats: unable to find data in memory cache]" Feb 14 12:26:05 crc kubenswrapper[4904]: I0214 12:26:05.837184 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:26:05 crc kubenswrapper[4904]: E0214 12:26:05.839029 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:26:06 crc kubenswrapper[4904]: E0214 12:26:06.752234 4904 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf278f66a_6395_4efb_bb4d_1628cd07d36c.slice\": RecentStats: unable to find data in memory cache]" Feb 14 12:26:17 crc kubenswrapper[4904]: I0214 12:26:17.837155 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:26:17 crc kubenswrapper[4904]: E0214 12:26:17.837813 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:26:31 crc kubenswrapper[4904]: I0214 12:26:31.836999 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:26:31 crc kubenswrapper[4904]: E0214 12:26:31.837744 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:26:44 crc kubenswrapper[4904]: I0214 12:26:44.837144 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:26:44 crc kubenswrapper[4904]: E0214 12:26:44.838212 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:26:55 crc kubenswrapper[4904]: I0214 12:26:55.837327 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:26:57 crc kubenswrapper[4904]: I0214 12:26:57.069091 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"c79f60f3a3c1cb61784ba260135e84fa21d0779eb625031dd8470a43f1ee7ee4"} Feb 14 12:28:20 crc kubenswrapper[4904]: I0214 12:28:20.208931 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x9rh9"] Feb 14 12:28:20 crc kubenswrapper[4904]: E0214 12:28:20.209964 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f278f66a-6395-4efb-bb4d-1628cd07d36c" containerName="extract-utilities" Feb 14 12:28:20 crc kubenswrapper[4904]: I0214 12:28:20.209979 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="f278f66a-6395-4efb-bb4d-1628cd07d36c" containerName="extract-utilities" Feb 14 12:28:20 crc kubenswrapper[4904]: E0214 12:28:20.210008 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f278f66a-6395-4efb-bb4d-1628cd07d36c" containerName="extract-content" Feb 14 12:28:20 crc kubenswrapper[4904]: I0214 12:28:20.210016 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="f278f66a-6395-4efb-bb4d-1628cd07d36c" containerName="extract-content" Feb 14 12:28:20 crc kubenswrapper[4904]: E0214 12:28:20.210026 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f278f66a-6395-4efb-bb4d-1628cd07d36c" containerName="registry-server" Feb 14 12:28:20 crc kubenswrapper[4904]: I0214 12:28:20.210053 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="f278f66a-6395-4efb-bb4d-1628cd07d36c" containerName="registry-server" Feb 14 12:28:20 crc kubenswrapper[4904]: I0214 12:28:20.210280 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="f278f66a-6395-4efb-bb4d-1628cd07d36c" containerName="registry-server" Feb 14 12:28:20 crc kubenswrapper[4904]: I0214 12:28:20.213624 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9rh9" Feb 14 12:28:20 crc kubenswrapper[4904]: I0214 12:28:20.229674 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x9rh9"] Feb 14 12:28:20 crc kubenswrapper[4904]: I0214 12:28:20.255787 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ppk2\" (UniqueName: \"kubernetes.io/projected/c21495d3-8411-4e89-97a1-49020a036cc9-kube-api-access-9ppk2\") pod \"redhat-operators-x9rh9\" (UID: \"c21495d3-8411-4e89-97a1-49020a036cc9\") " pod="openshift-marketplace/redhat-operators-x9rh9" Feb 14 12:28:20 crc kubenswrapper[4904]: I0214 12:28:20.256288 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c21495d3-8411-4e89-97a1-49020a036cc9-utilities\") pod \"redhat-operators-x9rh9\" (UID: \"c21495d3-8411-4e89-97a1-49020a036cc9\") " pod="openshift-marketplace/redhat-operators-x9rh9" Feb 14 12:28:20 crc kubenswrapper[4904]: I0214 12:28:20.256414 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c21495d3-8411-4e89-97a1-49020a036cc9-catalog-content\") pod \"redhat-operators-x9rh9\" (UID: \"c21495d3-8411-4e89-97a1-49020a036cc9\") " pod="openshift-marketplace/redhat-operators-x9rh9" Feb 14 12:28:20 crc kubenswrapper[4904]: I0214 12:28:20.357162 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c21495d3-8411-4e89-97a1-49020a036cc9-utilities\") pod \"redhat-operators-x9rh9\" (UID: \"c21495d3-8411-4e89-97a1-49020a036cc9\") " pod="openshift-marketplace/redhat-operators-x9rh9" Feb 14 12:28:20 crc kubenswrapper[4904]: I0214 12:28:20.357199 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c21495d3-8411-4e89-97a1-49020a036cc9-catalog-content\") pod \"redhat-operators-x9rh9\" (UID: \"c21495d3-8411-4e89-97a1-49020a036cc9\") " pod="openshift-marketplace/redhat-operators-x9rh9" Feb 14 12:28:20 crc kubenswrapper[4904]: I0214 12:28:20.357232 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ppk2\" (UniqueName: \"kubernetes.io/projected/c21495d3-8411-4e89-97a1-49020a036cc9-kube-api-access-9ppk2\") pod \"redhat-operators-x9rh9\" (UID: \"c21495d3-8411-4e89-97a1-49020a036cc9\") " pod="openshift-marketplace/redhat-operators-x9rh9" Feb 14 12:28:20 crc kubenswrapper[4904]: I0214 12:28:20.358211 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c21495d3-8411-4e89-97a1-49020a036cc9-utilities\") pod \"redhat-operators-x9rh9\" (UID: \"c21495d3-8411-4e89-97a1-49020a036cc9\") " pod="openshift-marketplace/redhat-operators-x9rh9" Feb 14 12:28:20 crc kubenswrapper[4904]: I0214 12:28:20.358277 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c21495d3-8411-4e89-97a1-49020a036cc9-catalog-content\") pod \"redhat-operators-x9rh9\" (UID: \"c21495d3-8411-4e89-97a1-49020a036cc9\") " pod="openshift-marketplace/redhat-operators-x9rh9" Feb 14 12:28:20 crc kubenswrapper[4904]: I0214 12:28:20.380589 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ppk2\" (UniqueName: \"kubernetes.io/projected/c21495d3-8411-4e89-97a1-49020a036cc9-kube-api-access-9ppk2\") pod \"redhat-operators-x9rh9\" (UID: \"c21495d3-8411-4e89-97a1-49020a036cc9\") " pod="openshift-marketplace/redhat-operators-x9rh9" Feb 14 12:28:20 crc kubenswrapper[4904]: I0214 12:28:20.537164 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9rh9" Feb 14 12:28:21 crc kubenswrapper[4904]: I0214 12:28:21.472163 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x9rh9"] Feb 14 12:28:21 crc kubenswrapper[4904]: I0214 12:28:21.726724 4904 generic.go:334] "Generic (PLEG): container finished" podID="c21495d3-8411-4e89-97a1-49020a036cc9" containerID="5cbf26cee5612bc2a39fb3e0f35b080469bf8f4f79dc3b4308adf2f21da8c929" exitCode=0 Feb 14 12:28:21 crc kubenswrapper[4904]: I0214 12:28:21.726810 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9rh9" event={"ID":"c21495d3-8411-4e89-97a1-49020a036cc9","Type":"ContainerDied","Data":"5cbf26cee5612bc2a39fb3e0f35b080469bf8f4f79dc3b4308adf2f21da8c929"} Feb 14 12:28:21 crc kubenswrapper[4904]: I0214 12:28:21.727038 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9rh9" event={"ID":"c21495d3-8411-4e89-97a1-49020a036cc9","Type":"ContainerStarted","Data":"cfb3f6dbb1da10f8a085975786d259215a23b840de13991d2645c4ec10afa83e"} Feb 14 12:28:23 crc kubenswrapper[4904]: I0214 12:28:23.752308 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9rh9" event={"ID":"c21495d3-8411-4e89-97a1-49020a036cc9","Type":"ContainerStarted","Data":"3836e78f45f13e3bc4ff2586665f4a96dfdb909ced6f26868b3285dbecaf73b9"} Feb 14 12:28:27 crc kubenswrapper[4904]: I0214 12:28:27.781791 4904 generic.go:334] "Generic (PLEG): container finished" podID="c21495d3-8411-4e89-97a1-49020a036cc9" containerID="3836e78f45f13e3bc4ff2586665f4a96dfdb909ced6f26868b3285dbecaf73b9" exitCode=0 Feb 14 12:28:27 crc kubenswrapper[4904]: I0214 12:28:27.782381 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9rh9" event={"ID":"c21495d3-8411-4e89-97a1-49020a036cc9","Type":"ContainerDied","Data":"3836e78f45f13e3bc4ff2586665f4a96dfdb909ced6f26868b3285dbecaf73b9"} Feb 14 12:28:28 crc kubenswrapper[4904]: I0214 12:28:28.792963 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9rh9" event={"ID":"c21495d3-8411-4e89-97a1-49020a036cc9","Type":"ContainerStarted","Data":"71a7cf39048a3ed0119fd5d888e8a4f5367914bb5231804f0b42df4631ceb78e"} Feb 14 12:28:28 crc kubenswrapper[4904]: I0214 12:28:28.813876 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x9rh9" podStartSLOduration=2.360401288 podStartE2EDuration="8.813857439s" podCreationTimestamp="2026-02-14 12:28:20 +0000 UTC" firstStartedPulling="2026-02-14 12:28:21.728362795 +0000 UTC m=+4692.541127456" lastFinishedPulling="2026-02-14 12:28:28.181818946 +0000 UTC m=+4698.994583607" observedRunningTime="2026-02-14 12:28:28.806604311 +0000 UTC m=+4699.619368972" watchObservedRunningTime="2026-02-14 12:28:28.813857439 +0000 UTC m=+4699.626622100" Feb 14 12:28:30 crc kubenswrapper[4904]: I0214 12:28:30.537513 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x9rh9" Feb 14 12:28:30 crc kubenswrapper[4904]: I0214 12:28:30.538083 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x9rh9" Feb 14 12:28:31 crc kubenswrapper[4904]: I0214 12:28:31.636007 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x9rh9" podUID="c21495d3-8411-4e89-97a1-49020a036cc9" containerName="registry-server" probeResult="failure" output=< Feb 14 12:28:31 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 12:28:31 crc kubenswrapper[4904]: > Feb 14 12:28:40 crc kubenswrapper[4904]: I0214 12:28:40.585906 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x9rh9" Feb 14 12:28:40 crc kubenswrapper[4904]: I0214 12:28:40.648922 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x9rh9" Feb 14 12:28:40 crc kubenswrapper[4904]: I0214 12:28:40.828206 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x9rh9"] Feb 14 12:28:41 crc kubenswrapper[4904]: I0214 12:28:41.903373 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x9rh9" podUID="c21495d3-8411-4e89-97a1-49020a036cc9" containerName="registry-server" containerID="cri-o://71a7cf39048a3ed0119fd5d888e8a4f5367914bb5231804f0b42df4631ceb78e" gracePeriod=2 Feb 14 12:28:42 crc kubenswrapper[4904]: I0214 12:28:42.394421 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9rh9" Feb 14 12:28:42 crc kubenswrapper[4904]: I0214 12:28:42.573209 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c21495d3-8411-4e89-97a1-49020a036cc9-utilities\") pod \"c21495d3-8411-4e89-97a1-49020a036cc9\" (UID: \"c21495d3-8411-4e89-97a1-49020a036cc9\") " Feb 14 12:28:42 crc kubenswrapper[4904]: I0214 12:28:42.573283 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ppk2\" (UniqueName: \"kubernetes.io/projected/c21495d3-8411-4e89-97a1-49020a036cc9-kube-api-access-9ppk2\") pod \"c21495d3-8411-4e89-97a1-49020a036cc9\" (UID: \"c21495d3-8411-4e89-97a1-49020a036cc9\") " Feb 14 12:28:42 crc kubenswrapper[4904]: I0214 12:28:42.573310 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c21495d3-8411-4e89-97a1-49020a036cc9-catalog-content\") pod \"c21495d3-8411-4e89-97a1-49020a036cc9\" (UID: \"c21495d3-8411-4e89-97a1-49020a036cc9\") " Feb 14 12:28:42 crc kubenswrapper[4904]: I0214 12:28:42.577770 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c21495d3-8411-4e89-97a1-49020a036cc9-utilities" (OuterVolumeSpecName: "utilities") pod "c21495d3-8411-4e89-97a1-49020a036cc9" (UID: "c21495d3-8411-4e89-97a1-49020a036cc9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:28:42 crc kubenswrapper[4904]: I0214 12:28:42.581035 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c21495d3-8411-4e89-97a1-49020a036cc9-kube-api-access-9ppk2" (OuterVolumeSpecName: "kube-api-access-9ppk2") pod "c21495d3-8411-4e89-97a1-49020a036cc9" (UID: "c21495d3-8411-4e89-97a1-49020a036cc9"). InnerVolumeSpecName "kube-api-access-9ppk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:28:42 crc kubenswrapper[4904]: I0214 12:28:42.675557 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c21495d3-8411-4e89-97a1-49020a036cc9-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 12:28:42 crc kubenswrapper[4904]: I0214 12:28:42.675825 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ppk2\" (UniqueName: \"kubernetes.io/projected/c21495d3-8411-4e89-97a1-49020a036cc9-kube-api-access-9ppk2\") on node \"crc\" DevicePath \"\"" Feb 14 12:28:42 crc kubenswrapper[4904]: I0214 12:28:42.711496 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c21495d3-8411-4e89-97a1-49020a036cc9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c21495d3-8411-4e89-97a1-49020a036cc9" (UID: "c21495d3-8411-4e89-97a1-49020a036cc9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:28:42 crc kubenswrapper[4904]: I0214 12:28:42.778060 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c21495d3-8411-4e89-97a1-49020a036cc9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 12:28:42 crc kubenswrapper[4904]: I0214 12:28:42.912876 4904 generic.go:334] "Generic (PLEG): container finished" podID="c21495d3-8411-4e89-97a1-49020a036cc9" containerID="71a7cf39048a3ed0119fd5d888e8a4f5367914bb5231804f0b42df4631ceb78e" exitCode=0 Feb 14 12:28:42 crc kubenswrapper[4904]: I0214 12:28:42.912925 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9rh9" Feb 14 12:28:42 crc kubenswrapper[4904]: I0214 12:28:42.912925 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9rh9" event={"ID":"c21495d3-8411-4e89-97a1-49020a036cc9","Type":"ContainerDied","Data":"71a7cf39048a3ed0119fd5d888e8a4f5367914bb5231804f0b42df4631ceb78e"} Feb 14 12:28:42 crc kubenswrapper[4904]: I0214 12:28:42.914925 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9rh9" event={"ID":"c21495d3-8411-4e89-97a1-49020a036cc9","Type":"ContainerDied","Data":"cfb3f6dbb1da10f8a085975786d259215a23b840de13991d2645c4ec10afa83e"} Feb 14 12:28:42 crc kubenswrapper[4904]: I0214 12:28:42.914961 4904 scope.go:117] "RemoveContainer" containerID="71a7cf39048a3ed0119fd5d888e8a4f5367914bb5231804f0b42df4631ceb78e" Feb 14 12:28:42 crc kubenswrapper[4904]: I0214 12:28:42.946516 4904 scope.go:117] "RemoveContainer" containerID="3836e78f45f13e3bc4ff2586665f4a96dfdb909ced6f26868b3285dbecaf73b9" Feb 14 12:28:42 crc kubenswrapper[4904]: I0214 12:28:42.980228 4904 scope.go:117] "RemoveContainer" containerID="5cbf26cee5612bc2a39fb3e0f35b080469bf8f4f79dc3b4308adf2f21da8c929" Feb 14 12:28:42 crc kubenswrapper[4904]: I0214 12:28:42.984646 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x9rh9"] Feb 14 12:28:43 crc kubenswrapper[4904]: I0214 12:28:43.003281 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x9rh9"] Feb 14 12:28:43 crc kubenswrapper[4904]: I0214 12:28:43.023942 4904 scope.go:117] "RemoveContainer" containerID="71a7cf39048a3ed0119fd5d888e8a4f5367914bb5231804f0b42df4631ceb78e" Feb 14 12:28:43 crc kubenswrapper[4904]: E0214 12:28:43.024276 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71a7cf39048a3ed0119fd5d888e8a4f5367914bb5231804f0b42df4631ceb78e\": container with ID starting with 71a7cf39048a3ed0119fd5d888e8a4f5367914bb5231804f0b42df4631ceb78e not found: ID does not exist" containerID="71a7cf39048a3ed0119fd5d888e8a4f5367914bb5231804f0b42df4631ceb78e" Feb 14 12:28:43 crc kubenswrapper[4904]: I0214 12:28:43.024317 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71a7cf39048a3ed0119fd5d888e8a4f5367914bb5231804f0b42df4631ceb78e"} err="failed to get container status \"71a7cf39048a3ed0119fd5d888e8a4f5367914bb5231804f0b42df4631ceb78e\": rpc error: code = NotFound desc = could not find container \"71a7cf39048a3ed0119fd5d888e8a4f5367914bb5231804f0b42df4631ceb78e\": container with ID starting with 71a7cf39048a3ed0119fd5d888e8a4f5367914bb5231804f0b42df4631ceb78e not found: ID does not exist" Feb 14 12:28:43 crc kubenswrapper[4904]: I0214 12:28:43.024342 4904 scope.go:117] "RemoveContainer" containerID="3836e78f45f13e3bc4ff2586665f4a96dfdb909ced6f26868b3285dbecaf73b9" Feb 14 12:28:43 crc kubenswrapper[4904]: E0214 12:28:43.024526 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3836e78f45f13e3bc4ff2586665f4a96dfdb909ced6f26868b3285dbecaf73b9\": container with ID starting with 3836e78f45f13e3bc4ff2586665f4a96dfdb909ced6f26868b3285dbecaf73b9 not found: ID does not exist" containerID="3836e78f45f13e3bc4ff2586665f4a96dfdb909ced6f26868b3285dbecaf73b9" Feb 14 12:28:43 crc kubenswrapper[4904]: I0214 12:28:43.024553 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3836e78f45f13e3bc4ff2586665f4a96dfdb909ced6f26868b3285dbecaf73b9"} err="failed to get container status \"3836e78f45f13e3bc4ff2586665f4a96dfdb909ced6f26868b3285dbecaf73b9\": rpc error: code = NotFound desc = could not find container \"3836e78f45f13e3bc4ff2586665f4a96dfdb909ced6f26868b3285dbecaf73b9\": container with ID starting with 3836e78f45f13e3bc4ff2586665f4a96dfdb909ced6f26868b3285dbecaf73b9 not found: ID does not exist" Feb 14 12:28:43 crc kubenswrapper[4904]: I0214 12:28:43.024566 4904 scope.go:117] "RemoveContainer" containerID="5cbf26cee5612bc2a39fb3e0f35b080469bf8f4f79dc3b4308adf2f21da8c929" Feb 14 12:28:43 crc kubenswrapper[4904]: E0214 12:28:43.024719 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cbf26cee5612bc2a39fb3e0f35b080469bf8f4f79dc3b4308adf2f21da8c929\": container with ID starting with 5cbf26cee5612bc2a39fb3e0f35b080469bf8f4f79dc3b4308adf2f21da8c929 not found: ID does not exist" containerID="5cbf26cee5612bc2a39fb3e0f35b080469bf8f4f79dc3b4308adf2f21da8c929" Feb 14 12:28:43 crc kubenswrapper[4904]: I0214 12:28:43.024737 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cbf26cee5612bc2a39fb3e0f35b080469bf8f4f79dc3b4308adf2f21da8c929"} err="failed to get container status \"5cbf26cee5612bc2a39fb3e0f35b080469bf8f4f79dc3b4308adf2f21da8c929\": rpc error: code = NotFound desc = could not find container \"5cbf26cee5612bc2a39fb3e0f35b080469bf8f4f79dc3b4308adf2f21da8c929\": container with ID starting with 5cbf26cee5612bc2a39fb3e0f35b080469bf8f4f79dc3b4308adf2f21da8c929 not found: ID does not exist" Feb 14 12:28:43 crc kubenswrapper[4904]: I0214 12:28:43.870171 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c21495d3-8411-4e89-97a1-49020a036cc9" path="/var/lib/kubelet/pods/c21495d3-8411-4e89-97a1-49020a036cc9/volumes" Feb 14 12:29:16 crc kubenswrapper[4904]: I0214 12:29:16.382934 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:29:16 crc kubenswrapper[4904]: I0214 12:29:16.384609 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:29:46 crc kubenswrapper[4904]: I0214 12:29:46.382749 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:29:46 crc kubenswrapper[4904]: I0214 12:29:46.383367 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.178491 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517870-w7sm4"] Feb 14 12:30:00 crc kubenswrapper[4904]: E0214 12:30:00.179411 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c21495d3-8411-4e89-97a1-49020a036cc9" containerName="registry-server" Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.179424 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="c21495d3-8411-4e89-97a1-49020a036cc9" containerName="registry-server" Feb 14 12:30:00 crc kubenswrapper[4904]: E0214 12:30:00.179448 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c21495d3-8411-4e89-97a1-49020a036cc9" containerName="extract-content" Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.179454 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="c21495d3-8411-4e89-97a1-49020a036cc9" containerName="extract-content" Feb 14 12:30:00 crc kubenswrapper[4904]: E0214 12:30:00.179469 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c21495d3-8411-4e89-97a1-49020a036cc9" containerName="extract-utilities" Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.179476 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="c21495d3-8411-4e89-97a1-49020a036cc9" containerName="extract-utilities" Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.179691 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="c21495d3-8411-4e89-97a1-49020a036cc9" containerName="registry-server" Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.180415 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517870-w7sm4" Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.185317 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.185510 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.259551 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517870-w7sm4"] Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.370242 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-972nt\" (UniqueName: \"kubernetes.io/projected/ca62cb45-75fd-401a-a936-a18423dba07f-kube-api-access-972nt\") pod \"collect-profiles-29517870-w7sm4\" (UID: \"ca62cb45-75fd-401a-a936-a18423dba07f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517870-w7sm4" Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.370300 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ca62cb45-75fd-401a-a936-a18423dba07f-config-volume\") pod \"collect-profiles-29517870-w7sm4\" (UID: \"ca62cb45-75fd-401a-a936-a18423dba07f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517870-w7sm4" Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.370364 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ca62cb45-75fd-401a-a936-a18423dba07f-secret-volume\") pod \"collect-profiles-29517870-w7sm4\" (UID: \"ca62cb45-75fd-401a-a936-a18423dba07f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517870-w7sm4" Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.471822 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-972nt\" (UniqueName: \"kubernetes.io/projected/ca62cb45-75fd-401a-a936-a18423dba07f-kube-api-access-972nt\") pod \"collect-profiles-29517870-w7sm4\" (UID: \"ca62cb45-75fd-401a-a936-a18423dba07f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517870-w7sm4" Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.471897 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ca62cb45-75fd-401a-a936-a18423dba07f-config-volume\") pod \"collect-profiles-29517870-w7sm4\" (UID: \"ca62cb45-75fd-401a-a936-a18423dba07f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517870-w7sm4" Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.471956 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ca62cb45-75fd-401a-a936-a18423dba07f-secret-volume\") pod \"collect-profiles-29517870-w7sm4\" (UID: \"ca62cb45-75fd-401a-a936-a18423dba07f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517870-w7sm4" Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.473268 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ca62cb45-75fd-401a-a936-a18423dba07f-config-volume\") pod \"collect-profiles-29517870-w7sm4\" (UID: \"ca62cb45-75fd-401a-a936-a18423dba07f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517870-w7sm4" Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.478812 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ca62cb45-75fd-401a-a936-a18423dba07f-secret-volume\") pod \"collect-profiles-29517870-w7sm4\" (UID: \"ca62cb45-75fd-401a-a936-a18423dba07f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517870-w7sm4" Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.490863 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-972nt\" (UniqueName: \"kubernetes.io/projected/ca62cb45-75fd-401a-a936-a18423dba07f-kube-api-access-972nt\") pod \"collect-profiles-29517870-w7sm4\" (UID: \"ca62cb45-75fd-401a-a936-a18423dba07f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29517870-w7sm4" Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.511821 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517870-w7sm4" Feb 14 12:30:00 crc kubenswrapper[4904]: I0214 12:30:00.937917 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517870-w7sm4"] Feb 14 12:30:01 crc kubenswrapper[4904]: I0214 12:30:01.658179 4904 generic.go:334] "Generic (PLEG): container finished" podID="ca62cb45-75fd-401a-a936-a18423dba07f" containerID="7dc6b88c4a14bfc96379b8b1583b3d353b18b0d496ccbc1f2902d98c40f434ca" exitCode=0 Feb 14 12:30:01 crc kubenswrapper[4904]: I0214 12:30:01.658394 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517870-w7sm4" event={"ID":"ca62cb45-75fd-401a-a936-a18423dba07f","Type":"ContainerDied","Data":"7dc6b88c4a14bfc96379b8b1583b3d353b18b0d496ccbc1f2902d98c40f434ca"} Feb 14 12:30:01 crc kubenswrapper[4904]: I0214 12:30:01.658420 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517870-w7sm4" event={"ID":"ca62cb45-75fd-401a-a936-a18423dba07f","Type":"ContainerStarted","Data":"964621cfee97cba669459f22680a2ccd35ec7c5a5f713f77e4dd429ae1bbfff2"} Feb 14 12:30:03 crc kubenswrapper[4904]: I0214 12:30:03.074042 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517870-w7sm4" Feb 14 12:30:03 crc kubenswrapper[4904]: I0214 12:30:03.221986 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ca62cb45-75fd-401a-a936-a18423dba07f-config-volume\") pod \"ca62cb45-75fd-401a-a936-a18423dba07f\" (UID: \"ca62cb45-75fd-401a-a936-a18423dba07f\") " Feb 14 12:30:03 crc kubenswrapper[4904]: I0214 12:30:03.222163 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ca62cb45-75fd-401a-a936-a18423dba07f-secret-volume\") pod \"ca62cb45-75fd-401a-a936-a18423dba07f\" (UID: \"ca62cb45-75fd-401a-a936-a18423dba07f\") " Feb 14 12:30:03 crc kubenswrapper[4904]: I0214 12:30:03.222325 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-972nt\" (UniqueName: \"kubernetes.io/projected/ca62cb45-75fd-401a-a936-a18423dba07f-kube-api-access-972nt\") pod \"ca62cb45-75fd-401a-a936-a18423dba07f\" (UID: \"ca62cb45-75fd-401a-a936-a18423dba07f\") " Feb 14 12:30:03 crc kubenswrapper[4904]: I0214 12:30:03.223947 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca62cb45-75fd-401a-a936-a18423dba07f-config-volume" (OuterVolumeSpecName: "config-volume") pod "ca62cb45-75fd-401a-a936-a18423dba07f" (UID: "ca62cb45-75fd-401a-a936-a18423dba07f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 14 12:30:03 crc kubenswrapper[4904]: I0214 12:30:03.230784 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca62cb45-75fd-401a-a936-a18423dba07f-kube-api-access-972nt" (OuterVolumeSpecName: "kube-api-access-972nt") pod "ca62cb45-75fd-401a-a936-a18423dba07f" (UID: "ca62cb45-75fd-401a-a936-a18423dba07f"). InnerVolumeSpecName "kube-api-access-972nt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:30:03 crc kubenswrapper[4904]: I0214 12:30:03.230814 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca62cb45-75fd-401a-a936-a18423dba07f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ca62cb45-75fd-401a-a936-a18423dba07f" (UID: "ca62cb45-75fd-401a-a936-a18423dba07f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 14 12:30:03 crc kubenswrapper[4904]: I0214 12:30:03.324502 4904 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ca62cb45-75fd-401a-a936-a18423dba07f-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 14 12:30:03 crc kubenswrapper[4904]: I0214 12:30:03.324541 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-972nt\" (UniqueName: \"kubernetes.io/projected/ca62cb45-75fd-401a-a936-a18423dba07f-kube-api-access-972nt\") on node \"crc\" DevicePath \"\"" Feb 14 12:30:03 crc kubenswrapper[4904]: I0214 12:30:03.324557 4904 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ca62cb45-75fd-401a-a936-a18423dba07f-config-volume\") on node \"crc\" DevicePath \"\"" Feb 14 12:30:03 crc kubenswrapper[4904]: I0214 12:30:03.674941 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29517870-w7sm4" event={"ID":"ca62cb45-75fd-401a-a936-a18423dba07f","Type":"ContainerDied","Data":"964621cfee97cba669459f22680a2ccd35ec7c5a5f713f77e4dd429ae1bbfff2"} Feb 14 12:30:03 crc kubenswrapper[4904]: I0214 12:30:03.675585 4904 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="964621cfee97cba669459f22680a2ccd35ec7c5a5f713f77e4dd429ae1bbfff2" Feb 14 12:30:03 crc kubenswrapper[4904]: I0214 12:30:03.675035 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29517870-w7sm4" Feb 14 12:30:04 crc kubenswrapper[4904]: I0214 12:30:04.147562 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l"] Feb 14 12:30:04 crc kubenswrapper[4904]: I0214 12:30:04.161951 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29517825-wzv4l"] Feb 14 12:30:05 crc kubenswrapper[4904]: I0214 12:30:05.856336 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c62364cf-6155-4246-be2f-48bb500d9fb7" path="/var/lib/kubelet/pods/c62364cf-6155-4246-be2f-48bb500d9fb7/volumes" Feb 14 12:30:16 crc kubenswrapper[4904]: I0214 12:30:16.382511 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:30:16 crc kubenswrapper[4904]: I0214 12:30:16.383238 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:30:16 crc kubenswrapper[4904]: I0214 12:30:16.383292 4904 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 12:30:16 crc kubenswrapper[4904]: I0214 12:30:16.384025 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c79f60f3a3c1cb61784ba260135e84fa21d0779eb625031dd8470a43f1ee7ee4"} pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 14 12:30:16 crc kubenswrapper[4904]: I0214 12:30:16.384093 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" containerID="cri-o://c79f60f3a3c1cb61784ba260135e84fa21d0779eb625031dd8470a43f1ee7ee4" gracePeriod=600 Feb 14 12:30:16 crc kubenswrapper[4904]: I0214 12:30:16.789503 4904 generic.go:334] "Generic (PLEG): container finished" podID="7790d276-2506-4641-8c22-67e6441417f0" containerID="c79f60f3a3c1cb61784ba260135e84fa21d0779eb625031dd8470a43f1ee7ee4" exitCode=0 Feb 14 12:30:16 crc kubenswrapper[4904]: I0214 12:30:16.789591 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerDied","Data":"c79f60f3a3c1cb61784ba260135e84fa21d0779eb625031dd8470a43f1ee7ee4"} Feb 14 12:30:16 crc kubenswrapper[4904]: I0214 12:30:16.789913 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658"} Feb 14 12:30:16 crc kubenswrapper[4904]: I0214 12:30:16.789954 4904 scope.go:117] "RemoveContainer" containerID="fa974e5060d098fb51cec1ec9e27cf0d87d01a6fd3163dee9ccaeda5bf1b3f28" Feb 14 12:30:16 crc kubenswrapper[4904]: I0214 12:30:16.958792 4904 scope.go:117] "RemoveContainer" containerID="5733dc3b95e94b18e6c617fe41c7000f51016214880747bc947afa1962e680b8" Feb 14 12:30:43 crc kubenswrapper[4904]: I0214 12:30:43.209054 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fxrwd"] Feb 14 12:30:43 crc kubenswrapper[4904]: E0214 12:30:43.210052 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca62cb45-75fd-401a-a936-a18423dba07f" containerName="collect-profiles" Feb 14 12:30:43 crc kubenswrapper[4904]: I0214 12:30:43.210070 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca62cb45-75fd-401a-a936-a18423dba07f" containerName="collect-profiles" Feb 14 12:30:43 crc kubenswrapper[4904]: I0214 12:30:43.210323 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca62cb45-75fd-401a-a936-a18423dba07f" containerName="collect-profiles" Feb 14 12:30:43 crc kubenswrapper[4904]: I0214 12:30:43.212020 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fxrwd" Feb 14 12:30:43 crc kubenswrapper[4904]: I0214 12:30:43.217292 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fxrwd"] Feb 14 12:30:43 crc kubenswrapper[4904]: I0214 12:30:43.270688 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdw95\" (UniqueName: \"kubernetes.io/projected/8dcafde8-fb32-49b9-aadb-b1662f5ddb90-kube-api-access-bdw95\") pod \"redhat-marketplace-fxrwd\" (UID: \"8dcafde8-fb32-49b9-aadb-b1662f5ddb90\") " pod="openshift-marketplace/redhat-marketplace-fxrwd" Feb 14 12:30:43 crc kubenswrapper[4904]: I0214 12:30:43.270763 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dcafde8-fb32-49b9-aadb-b1662f5ddb90-catalog-content\") pod \"redhat-marketplace-fxrwd\" (UID: \"8dcafde8-fb32-49b9-aadb-b1662f5ddb90\") " pod="openshift-marketplace/redhat-marketplace-fxrwd" Feb 14 12:30:43 crc kubenswrapper[4904]: I0214 12:30:43.270862 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dcafde8-fb32-49b9-aadb-b1662f5ddb90-utilities\") pod \"redhat-marketplace-fxrwd\" (UID: \"8dcafde8-fb32-49b9-aadb-b1662f5ddb90\") " pod="openshift-marketplace/redhat-marketplace-fxrwd" Feb 14 12:30:43 crc kubenswrapper[4904]: I0214 12:30:43.372303 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdw95\" (UniqueName: \"kubernetes.io/projected/8dcafde8-fb32-49b9-aadb-b1662f5ddb90-kube-api-access-bdw95\") pod \"redhat-marketplace-fxrwd\" (UID: \"8dcafde8-fb32-49b9-aadb-b1662f5ddb90\") " pod="openshift-marketplace/redhat-marketplace-fxrwd" Feb 14 12:30:43 crc kubenswrapper[4904]: I0214 12:30:43.372644 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dcafde8-fb32-49b9-aadb-b1662f5ddb90-catalog-content\") pod \"redhat-marketplace-fxrwd\" (UID: \"8dcafde8-fb32-49b9-aadb-b1662f5ddb90\") " pod="openshift-marketplace/redhat-marketplace-fxrwd" Feb 14 12:30:43 crc kubenswrapper[4904]: I0214 12:30:43.372722 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dcafde8-fb32-49b9-aadb-b1662f5ddb90-utilities\") pod \"redhat-marketplace-fxrwd\" (UID: \"8dcafde8-fb32-49b9-aadb-b1662f5ddb90\") " pod="openshift-marketplace/redhat-marketplace-fxrwd" Feb 14 12:30:43 crc kubenswrapper[4904]: I0214 12:30:43.373200 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dcafde8-fb32-49b9-aadb-b1662f5ddb90-utilities\") pod \"redhat-marketplace-fxrwd\" (UID: \"8dcafde8-fb32-49b9-aadb-b1662f5ddb90\") " pod="openshift-marketplace/redhat-marketplace-fxrwd" Feb 14 12:30:43 crc kubenswrapper[4904]: I0214 12:30:43.373464 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dcafde8-fb32-49b9-aadb-b1662f5ddb90-catalog-content\") pod \"redhat-marketplace-fxrwd\" (UID: \"8dcafde8-fb32-49b9-aadb-b1662f5ddb90\") " pod="openshift-marketplace/redhat-marketplace-fxrwd" Feb 14 12:30:43 crc kubenswrapper[4904]: I0214 12:30:43.409463 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdw95\" (UniqueName: \"kubernetes.io/projected/8dcafde8-fb32-49b9-aadb-b1662f5ddb90-kube-api-access-bdw95\") pod \"redhat-marketplace-fxrwd\" (UID: \"8dcafde8-fb32-49b9-aadb-b1662f5ddb90\") " pod="openshift-marketplace/redhat-marketplace-fxrwd" Feb 14 12:30:43 crc kubenswrapper[4904]: I0214 12:30:43.544463 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fxrwd" Feb 14 12:30:44 crc kubenswrapper[4904]: I0214 12:30:44.024151 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fxrwd"] Feb 14 12:30:45 crc kubenswrapper[4904]: I0214 12:30:45.041224 4904 generic.go:334] "Generic (PLEG): container finished" podID="8dcafde8-fb32-49b9-aadb-b1662f5ddb90" containerID="0dbfe7066a145cf9714621f1cc38f4ae0b3131d14b4c9bbb5a755d642d02d175" exitCode=0 Feb 14 12:30:45 crc kubenswrapper[4904]: I0214 12:30:45.041339 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxrwd" event={"ID":"8dcafde8-fb32-49b9-aadb-b1662f5ddb90","Type":"ContainerDied","Data":"0dbfe7066a145cf9714621f1cc38f4ae0b3131d14b4c9bbb5a755d642d02d175"} Feb 14 12:30:45 crc kubenswrapper[4904]: I0214 12:30:45.041559 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxrwd" event={"ID":"8dcafde8-fb32-49b9-aadb-b1662f5ddb90","Type":"ContainerStarted","Data":"6bd83f5611ed844ab699bacb505d058866156ddb3f99f15a6d306000d706cd0b"} Feb 14 12:30:45 crc kubenswrapper[4904]: I0214 12:30:45.043764 4904 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 14 12:30:46 crc kubenswrapper[4904]: I0214 12:30:46.050874 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxrwd" event={"ID":"8dcafde8-fb32-49b9-aadb-b1662f5ddb90","Type":"ContainerStarted","Data":"7576762daa63191d90d44b369733b9140926cc36243b191e4f5c48bc09e10e43"} Feb 14 12:30:47 crc kubenswrapper[4904]: I0214 12:30:47.064254 4904 generic.go:334] "Generic (PLEG): container finished" podID="8dcafde8-fb32-49b9-aadb-b1662f5ddb90" containerID="7576762daa63191d90d44b369733b9140926cc36243b191e4f5c48bc09e10e43" exitCode=0 Feb 14 12:30:47 crc kubenswrapper[4904]: I0214 12:30:47.064391 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxrwd" event={"ID":"8dcafde8-fb32-49b9-aadb-b1662f5ddb90","Type":"ContainerDied","Data":"7576762daa63191d90d44b369733b9140926cc36243b191e4f5c48bc09e10e43"} Feb 14 12:30:48 crc kubenswrapper[4904]: I0214 12:30:48.072802 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxrwd" event={"ID":"8dcafde8-fb32-49b9-aadb-b1662f5ddb90","Type":"ContainerStarted","Data":"f441f0f1adbb1ba736b8acdb4591d88e73c0a396a88ac892960dd1a0e2902ccf"} Feb 14 12:30:48 crc kubenswrapper[4904]: I0214 12:30:48.090464 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fxrwd" podStartSLOduration=2.445936583 podStartE2EDuration="5.0904489s" podCreationTimestamp="2026-02-14 12:30:43 +0000 UTC" firstStartedPulling="2026-02-14 12:30:45.043538518 +0000 UTC m=+4835.856303179" lastFinishedPulling="2026-02-14 12:30:47.688050825 +0000 UTC m=+4838.500815496" observedRunningTime="2026-02-14 12:30:48.086962965 +0000 UTC m=+4838.899727636" watchObservedRunningTime="2026-02-14 12:30:48.0904489 +0000 UTC m=+4838.903213561" Feb 14 12:30:53 crc kubenswrapper[4904]: I0214 12:30:53.545953 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fxrwd" Feb 14 12:30:53 crc kubenswrapper[4904]: I0214 12:30:53.546698 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fxrwd" Feb 14 12:30:53 crc kubenswrapper[4904]: I0214 12:30:53.612055 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fxrwd" Feb 14 12:30:54 crc kubenswrapper[4904]: I0214 12:30:54.157068 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fxrwd" Feb 14 12:30:54 crc kubenswrapper[4904]: I0214 12:30:54.198576 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fxrwd"] Feb 14 12:30:56 crc kubenswrapper[4904]: I0214 12:30:56.132070 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fxrwd" podUID="8dcafde8-fb32-49b9-aadb-b1662f5ddb90" containerName="registry-server" containerID="cri-o://f441f0f1adbb1ba736b8acdb4591d88e73c0a396a88ac892960dd1a0e2902ccf" gracePeriod=2 Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.066608 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fxrwd" Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.147396 4904 generic.go:334] "Generic (PLEG): container finished" podID="8dcafde8-fb32-49b9-aadb-b1662f5ddb90" containerID="f441f0f1adbb1ba736b8acdb4591d88e73c0a396a88ac892960dd1a0e2902ccf" exitCode=0 Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.147440 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxrwd" event={"ID":"8dcafde8-fb32-49b9-aadb-b1662f5ddb90","Type":"ContainerDied","Data":"f441f0f1adbb1ba736b8acdb4591d88e73c0a396a88ac892960dd1a0e2902ccf"} Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.147494 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxrwd" event={"ID":"8dcafde8-fb32-49b9-aadb-b1662f5ddb90","Type":"ContainerDied","Data":"6bd83f5611ed844ab699bacb505d058866156ddb3f99f15a6d306000d706cd0b"} Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.147519 4904 scope.go:117] "RemoveContainer" containerID="f441f0f1adbb1ba736b8acdb4591d88e73c0a396a88ac892960dd1a0e2902ccf" Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.147515 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fxrwd" Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.175015 4904 scope.go:117] "RemoveContainer" containerID="7576762daa63191d90d44b369733b9140926cc36243b191e4f5c48bc09e10e43" Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.204015 4904 scope.go:117] "RemoveContainer" containerID="0dbfe7066a145cf9714621f1cc38f4ae0b3131d14b4c9bbb5a755d642d02d175" Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.226000 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdw95\" (UniqueName: \"kubernetes.io/projected/8dcafde8-fb32-49b9-aadb-b1662f5ddb90-kube-api-access-bdw95\") pod \"8dcafde8-fb32-49b9-aadb-b1662f5ddb90\" (UID: \"8dcafde8-fb32-49b9-aadb-b1662f5ddb90\") " Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.226058 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dcafde8-fb32-49b9-aadb-b1662f5ddb90-catalog-content\") pod \"8dcafde8-fb32-49b9-aadb-b1662f5ddb90\" (UID: \"8dcafde8-fb32-49b9-aadb-b1662f5ddb90\") " Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.226225 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dcafde8-fb32-49b9-aadb-b1662f5ddb90-utilities\") pod \"8dcafde8-fb32-49b9-aadb-b1662f5ddb90\" (UID: \"8dcafde8-fb32-49b9-aadb-b1662f5ddb90\") " Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.227279 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dcafde8-fb32-49b9-aadb-b1662f5ddb90-utilities" (OuterVolumeSpecName: "utilities") pod "8dcafde8-fb32-49b9-aadb-b1662f5ddb90" (UID: "8dcafde8-fb32-49b9-aadb-b1662f5ddb90"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.232514 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dcafde8-fb32-49b9-aadb-b1662f5ddb90-kube-api-access-bdw95" (OuterVolumeSpecName: "kube-api-access-bdw95") pod "8dcafde8-fb32-49b9-aadb-b1662f5ddb90" (UID: "8dcafde8-fb32-49b9-aadb-b1662f5ddb90"). InnerVolumeSpecName "kube-api-access-bdw95". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.236925 4904 scope.go:117] "RemoveContainer" containerID="f441f0f1adbb1ba736b8acdb4591d88e73c0a396a88ac892960dd1a0e2902ccf" Feb 14 12:30:57 crc kubenswrapper[4904]: E0214 12:30:57.237346 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f441f0f1adbb1ba736b8acdb4591d88e73c0a396a88ac892960dd1a0e2902ccf\": container with ID starting with f441f0f1adbb1ba736b8acdb4591d88e73c0a396a88ac892960dd1a0e2902ccf not found: ID does not exist" containerID="f441f0f1adbb1ba736b8acdb4591d88e73c0a396a88ac892960dd1a0e2902ccf" Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.237391 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f441f0f1adbb1ba736b8acdb4591d88e73c0a396a88ac892960dd1a0e2902ccf"} err="failed to get container status \"f441f0f1adbb1ba736b8acdb4591d88e73c0a396a88ac892960dd1a0e2902ccf\": rpc error: code = NotFound desc = could not find container \"f441f0f1adbb1ba736b8acdb4591d88e73c0a396a88ac892960dd1a0e2902ccf\": container with ID starting with f441f0f1adbb1ba736b8acdb4591d88e73c0a396a88ac892960dd1a0e2902ccf not found: ID does not exist" Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.237420 4904 scope.go:117] "RemoveContainer" containerID="7576762daa63191d90d44b369733b9140926cc36243b191e4f5c48bc09e10e43" Feb 14 12:30:57 crc kubenswrapper[4904]: E0214 12:30:57.237818 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7576762daa63191d90d44b369733b9140926cc36243b191e4f5c48bc09e10e43\": container with ID starting with 7576762daa63191d90d44b369733b9140926cc36243b191e4f5c48bc09e10e43 not found: ID does not exist" containerID="7576762daa63191d90d44b369733b9140926cc36243b191e4f5c48bc09e10e43" Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.238425 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7576762daa63191d90d44b369733b9140926cc36243b191e4f5c48bc09e10e43"} err="failed to get container status \"7576762daa63191d90d44b369733b9140926cc36243b191e4f5c48bc09e10e43\": rpc error: code = NotFound desc = could not find container \"7576762daa63191d90d44b369733b9140926cc36243b191e4f5c48bc09e10e43\": container with ID starting with 7576762daa63191d90d44b369733b9140926cc36243b191e4f5c48bc09e10e43 not found: ID does not exist" Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.238452 4904 scope.go:117] "RemoveContainer" containerID="0dbfe7066a145cf9714621f1cc38f4ae0b3131d14b4c9bbb5a755d642d02d175" Feb 14 12:30:57 crc kubenswrapper[4904]: E0214 12:30:57.238821 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dbfe7066a145cf9714621f1cc38f4ae0b3131d14b4c9bbb5a755d642d02d175\": container with ID starting with 0dbfe7066a145cf9714621f1cc38f4ae0b3131d14b4c9bbb5a755d642d02d175 not found: ID does not exist" containerID="0dbfe7066a145cf9714621f1cc38f4ae0b3131d14b4c9bbb5a755d642d02d175" Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.238854 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dbfe7066a145cf9714621f1cc38f4ae0b3131d14b4c9bbb5a755d642d02d175"} err="failed to get container status \"0dbfe7066a145cf9714621f1cc38f4ae0b3131d14b4c9bbb5a755d642d02d175\": rpc error: code = NotFound desc = could not find container \"0dbfe7066a145cf9714621f1cc38f4ae0b3131d14b4c9bbb5a755d642d02d175\": container with ID starting with 0dbfe7066a145cf9714621f1cc38f4ae0b3131d14b4c9bbb5a755d642d02d175 not found: ID does not exist" Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.252022 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dcafde8-fb32-49b9-aadb-b1662f5ddb90-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8dcafde8-fb32-49b9-aadb-b1662f5ddb90" (UID: "8dcafde8-fb32-49b9-aadb-b1662f5ddb90"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.328471 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dcafde8-fb32-49b9-aadb-b1662f5ddb90-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.328500 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdw95\" (UniqueName: \"kubernetes.io/projected/8dcafde8-fb32-49b9-aadb-b1662f5ddb90-kube-api-access-bdw95\") on node \"crc\" DevicePath \"\"" Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.328511 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dcafde8-fb32-49b9-aadb-b1662f5ddb90-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.484636 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fxrwd"] Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.493086 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fxrwd"] Feb 14 12:30:57 crc kubenswrapper[4904]: I0214 12:30:57.848575 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dcafde8-fb32-49b9-aadb-b1662f5ddb90" path="/var/lib/kubelet/pods/8dcafde8-fb32-49b9-aadb-b1662f5ddb90/volumes" Feb 14 12:31:09 crc kubenswrapper[4904]: I0214 12:31:09.468069 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-whwk6/must-gather-cdvfw"] Feb 14 12:31:09 crc kubenswrapper[4904]: E0214 12:31:09.469905 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcafde8-fb32-49b9-aadb-b1662f5ddb90" containerName="extract-content" Feb 14 12:31:09 crc kubenswrapper[4904]: I0214 12:31:09.470000 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcafde8-fb32-49b9-aadb-b1662f5ddb90" containerName="extract-content" Feb 14 12:31:09 crc kubenswrapper[4904]: E0214 12:31:09.472334 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcafde8-fb32-49b9-aadb-b1662f5ddb90" containerName="extract-utilities" Feb 14 12:31:09 crc kubenswrapper[4904]: I0214 12:31:09.472430 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcafde8-fb32-49b9-aadb-b1662f5ddb90" containerName="extract-utilities" Feb 14 12:31:09 crc kubenswrapper[4904]: E0214 12:31:09.472816 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcafde8-fb32-49b9-aadb-b1662f5ddb90" containerName="registry-server" Feb 14 12:31:09 crc kubenswrapper[4904]: I0214 12:31:09.472936 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcafde8-fb32-49b9-aadb-b1662f5ddb90" containerName="registry-server" Feb 14 12:31:09 crc kubenswrapper[4904]: I0214 12:31:09.473338 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dcafde8-fb32-49b9-aadb-b1662f5ddb90" containerName="registry-server" Feb 14 12:31:09 crc kubenswrapper[4904]: I0214 12:31:09.474775 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-whwk6/must-gather-cdvfw" Feb 14 12:31:09 crc kubenswrapper[4904]: I0214 12:31:09.480110 4904 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-whwk6"/"default-dockercfg-q6ktt" Feb 14 12:31:09 crc kubenswrapper[4904]: I0214 12:31:09.489187 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-whwk6"/"openshift-service-ca.crt" Feb 14 12:31:09 crc kubenswrapper[4904]: I0214 12:31:09.491054 4904 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-whwk6"/"kube-root-ca.crt" Feb 14 12:31:09 crc kubenswrapper[4904]: I0214 12:31:09.501362 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-whwk6/must-gather-cdvfw"] Feb 14 12:31:09 crc kubenswrapper[4904]: I0214 12:31:09.641024 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2283857f-6998-4d33-b42b-40c360586a64-must-gather-output\") pod \"must-gather-cdvfw\" (UID: \"2283857f-6998-4d33-b42b-40c360586a64\") " pod="openshift-must-gather-whwk6/must-gather-cdvfw" Feb 14 12:31:09 crc kubenswrapper[4904]: I0214 12:31:09.641142 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99bbv\" (UniqueName: \"kubernetes.io/projected/2283857f-6998-4d33-b42b-40c360586a64-kube-api-access-99bbv\") pod \"must-gather-cdvfw\" (UID: \"2283857f-6998-4d33-b42b-40c360586a64\") " pod="openshift-must-gather-whwk6/must-gather-cdvfw" Feb 14 12:31:09 crc kubenswrapper[4904]: I0214 12:31:09.742997 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2283857f-6998-4d33-b42b-40c360586a64-must-gather-output\") pod \"must-gather-cdvfw\" (UID: \"2283857f-6998-4d33-b42b-40c360586a64\") " pod="openshift-must-gather-whwk6/must-gather-cdvfw" Feb 14 12:31:09 crc kubenswrapper[4904]: I0214 12:31:09.743097 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99bbv\" (UniqueName: \"kubernetes.io/projected/2283857f-6998-4d33-b42b-40c360586a64-kube-api-access-99bbv\") pod \"must-gather-cdvfw\" (UID: \"2283857f-6998-4d33-b42b-40c360586a64\") " pod="openshift-must-gather-whwk6/must-gather-cdvfw" Feb 14 12:31:09 crc kubenswrapper[4904]: I0214 12:31:09.743799 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2283857f-6998-4d33-b42b-40c360586a64-must-gather-output\") pod \"must-gather-cdvfw\" (UID: \"2283857f-6998-4d33-b42b-40c360586a64\") " pod="openshift-must-gather-whwk6/must-gather-cdvfw" Feb 14 12:31:09 crc kubenswrapper[4904]: I0214 12:31:09.760956 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99bbv\" (UniqueName: \"kubernetes.io/projected/2283857f-6998-4d33-b42b-40c360586a64-kube-api-access-99bbv\") pod \"must-gather-cdvfw\" (UID: \"2283857f-6998-4d33-b42b-40c360586a64\") " pod="openshift-must-gather-whwk6/must-gather-cdvfw" Feb 14 12:31:09 crc kubenswrapper[4904]: I0214 12:31:09.791975 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-whwk6/must-gather-cdvfw" Feb 14 12:31:10 crc kubenswrapper[4904]: I0214 12:31:10.306785 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-whwk6/must-gather-cdvfw"] Feb 14 12:31:11 crc kubenswrapper[4904]: I0214 12:31:11.261145 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-whwk6/must-gather-cdvfw" event={"ID":"2283857f-6998-4d33-b42b-40c360586a64","Type":"ContainerStarted","Data":"8d4a987c0dc5f8ec9642cb66c8769dd64795f638bc65f0bb3fe0164a3d939b2e"} Feb 14 12:31:17 crc kubenswrapper[4904]: I0214 12:31:17.340720 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-whwk6/must-gather-cdvfw" event={"ID":"2283857f-6998-4d33-b42b-40c360586a64","Type":"ContainerStarted","Data":"e71823cf89cf36c0c1b4707d050d4ac82a47cdc22a3061e00c700a665312eee1"} Feb 14 12:31:18 crc kubenswrapper[4904]: I0214 12:31:18.350789 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-whwk6/must-gather-cdvfw" event={"ID":"2283857f-6998-4d33-b42b-40c360586a64","Type":"ContainerStarted","Data":"75dd94c86c2a31e73d5e0986164419e967c3c97b54727d6290e89687501680ad"} Feb 14 12:31:18 crc kubenswrapper[4904]: I0214 12:31:18.371197 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-whwk6/must-gather-cdvfw" podStartSLOduration=2.770030056 podStartE2EDuration="9.371170339s" podCreationTimestamp="2026-02-14 12:31:09 +0000 UTC" firstStartedPulling="2026-02-14 12:31:10.322514303 +0000 UTC m=+4861.135278974" lastFinishedPulling="2026-02-14 12:31:16.923654586 +0000 UTC m=+4867.736419257" observedRunningTime="2026-02-14 12:31:18.364857147 +0000 UTC m=+4869.177621818" watchObservedRunningTime="2026-02-14 12:31:18.371170339 +0000 UTC m=+4869.183935000" Feb 14 12:31:22 crc kubenswrapper[4904]: I0214 12:31:22.315101 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-whwk6/crc-debug-9qj7h"] Feb 14 12:31:22 crc kubenswrapper[4904]: I0214 12:31:22.323212 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-whwk6/crc-debug-9qj7h" Feb 14 12:31:22 crc kubenswrapper[4904]: I0214 12:31:22.386723 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2qq9\" (UniqueName: \"kubernetes.io/projected/43bb26db-1573-4324-bceb-a21e1ab4ceac-kube-api-access-t2qq9\") pod \"crc-debug-9qj7h\" (UID: \"43bb26db-1573-4324-bceb-a21e1ab4ceac\") " pod="openshift-must-gather-whwk6/crc-debug-9qj7h" Feb 14 12:31:22 crc kubenswrapper[4904]: I0214 12:31:22.386941 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/43bb26db-1573-4324-bceb-a21e1ab4ceac-host\") pod \"crc-debug-9qj7h\" (UID: \"43bb26db-1573-4324-bceb-a21e1ab4ceac\") " pod="openshift-must-gather-whwk6/crc-debug-9qj7h" Feb 14 12:31:22 crc kubenswrapper[4904]: I0214 12:31:22.488249 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/43bb26db-1573-4324-bceb-a21e1ab4ceac-host\") pod \"crc-debug-9qj7h\" (UID: \"43bb26db-1573-4324-bceb-a21e1ab4ceac\") " pod="openshift-must-gather-whwk6/crc-debug-9qj7h" Feb 14 12:31:22 crc kubenswrapper[4904]: I0214 12:31:22.488353 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2qq9\" (UniqueName: \"kubernetes.io/projected/43bb26db-1573-4324-bceb-a21e1ab4ceac-kube-api-access-t2qq9\") pod \"crc-debug-9qj7h\" (UID: \"43bb26db-1573-4324-bceb-a21e1ab4ceac\") " pod="openshift-must-gather-whwk6/crc-debug-9qj7h" Feb 14 12:31:22 crc kubenswrapper[4904]: I0214 12:31:22.488419 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/43bb26db-1573-4324-bceb-a21e1ab4ceac-host\") pod \"crc-debug-9qj7h\" (UID: \"43bb26db-1573-4324-bceb-a21e1ab4ceac\") " pod="openshift-must-gather-whwk6/crc-debug-9qj7h" Feb 14 12:31:22 crc kubenswrapper[4904]: I0214 12:31:22.518586 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2qq9\" (UniqueName: \"kubernetes.io/projected/43bb26db-1573-4324-bceb-a21e1ab4ceac-kube-api-access-t2qq9\") pod \"crc-debug-9qj7h\" (UID: \"43bb26db-1573-4324-bceb-a21e1ab4ceac\") " pod="openshift-must-gather-whwk6/crc-debug-9qj7h" Feb 14 12:31:22 crc kubenswrapper[4904]: I0214 12:31:22.646465 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-whwk6/crc-debug-9qj7h" Feb 14 12:31:23 crc kubenswrapper[4904]: I0214 12:31:23.400016 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-whwk6/crc-debug-9qj7h" event={"ID":"43bb26db-1573-4324-bceb-a21e1ab4ceac","Type":"ContainerStarted","Data":"cf03dcb7e8fece2e467b29a08d3340c08adaccda84f3114669a518f3c0918f72"} Feb 14 12:31:24 crc kubenswrapper[4904]: E0214 12:31:24.393179 4904 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.17:40464->38.102.83.17:43083: write tcp 38.102.83.17:40464->38.102.83.17:43083: write: broken pipe Feb 14 12:31:35 crc kubenswrapper[4904]: I0214 12:31:35.498909 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-whwk6/crc-debug-9qj7h" event={"ID":"43bb26db-1573-4324-bceb-a21e1ab4ceac","Type":"ContainerStarted","Data":"e635bddaaddba760061e86376a842cf53e498326309bf68e96a551e87f46c1f8"} Feb 14 12:31:35 crc kubenswrapper[4904]: I0214 12:31:35.520395 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-whwk6/crc-debug-9qj7h" podStartSLOduration=1.575249607 podStartE2EDuration="13.520378175s" podCreationTimestamp="2026-02-14 12:31:22 +0000 UTC" firstStartedPulling="2026-02-14 12:31:22.691210904 +0000 UTC m=+4873.503975565" lastFinishedPulling="2026-02-14 12:31:34.636339472 +0000 UTC m=+4885.449104133" observedRunningTime="2026-02-14 12:31:35.519926352 +0000 UTC m=+4886.332691013" watchObservedRunningTime="2026-02-14 12:31:35.520378175 +0000 UTC m=+4886.333142836" Feb 14 12:31:51 crc kubenswrapper[4904]: I0214 12:31:51.979377 4904 generic.go:334] "Generic (PLEG): container finished" podID="43bb26db-1573-4324-bceb-a21e1ab4ceac" containerID="e635bddaaddba760061e86376a842cf53e498326309bf68e96a551e87f46c1f8" exitCode=0 Feb 14 12:31:51 crc kubenswrapper[4904]: I0214 12:31:51.979563 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-whwk6/crc-debug-9qj7h" event={"ID":"43bb26db-1573-4324-bceb-a21e1ab4ceac","Type":"ContainerDied","Data":"e635bddaaddba760061e86376a842cf53e498326309bf68e96a551e87f46c1f8"} Feb 14 12:31:53 crc kubenswrapper[4904]: I0214 12:31:53.104133 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-whwk6/crc-debug-9qj7h" Feb 14 12:31:53 crc kubenswrapper[4904]: I0214 12:31:53.136942 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-whwk6/crc-debug-9qj7h"] Feb 14 12:31:53 crc kubenswrapper[4904]: I0214 12:31:53.146234 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-whwk6/crc-debug-9qj7h"] Feb 14 12:31:53 crc kubenswrapper[4904]: I0214 12:31:53.224663 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2qq9\" (UniqueName: \"kubernetes.io/projected/43bb26db-1573-4324-bceb-a21e1ab4ceac-kube-api-access-t2qq9\") pod \"43bb26db-1573-4324-bceb-a21e1ab4ceac\" (UID: \"43bb26db-1573-4324-bceb-a21e1ab4ceac\") " Feb 14 12:31:53 crc kubenswrapper[4904]: I0214 12:31:53.225003 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/43bb26db-1573-4324-bceb-a21e1ab4ceac-host\") pod \"43bb26db-1573-4324-bceb-a21e1ab4ceac\" (UID: \"43bb26db-1573-4324-bceb-a21e1ab4ceac\") " Feb 14 12:31:53 crc kubenswrapper[4904]: I0214 12:31:53.226424 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43bb26db-1573-4324-bceb-a21e1ab4ceac-host" (OuterVolumeSpecName: "host") pod "43bb26db-1573-4324-bceb-a21e1ab4ceac" (UID: "43bb26db-1573-4324-bceb-a21e1ab4ceac"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 12:31:53 crc kubenswrapper[4904]: I0214 12:31:53.236026 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43bb26db-1573-4324-bceb-a21e1ab4ceac-kube-api-access-t2qq9" (OuterVolumeSpecName: "kube-api-access-t2qq9") pod "43bb26db-1573-4324-bceb-a21e1ab4ceac" (UID: "43bb26db-1573-4324-bceb-a21e1ab4ceac"). InnerVolumeSpecName "kube-api-access-t2qq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:31:53 crc kubenswrapper[4904]: I0214 12:31:53.327818 4904 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/43bb26db-1573-4324-bceb-a21e1ab4ceac-host\") on node \"crc\" DevicePath \"\"" Feb 14 12:31:53 crc kubenswrapper[4904]: I0214 12:31:53.327869 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2qq9\" (UniqueName: \"kubernetes.io/projected/43bb26db-1573-4324-bceb-a21e1ab4ceac-kube-api-access-t2qq9\") on node \"crc\" DevicePath \"\"" Feb 14 12:31:53 crc kubenswrapper[4904]: I0214 12:31:53.849156 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43bb26db-1573-4324-bceb-a21e1ab4ceac" path="/var/lib/kubelet/pods/43bb26db-1573-4324-bceb-a21e1ab4ceac/volumes" Feb 14 12:31:53 crc kubenswrapper[4904]: I0214 12:31:53.994748 4904 scope.go:117] "RemoveContainer" containerID="e635bddaaddba760061e86376a842cf53e498326309bf68e96a551e87f46c1f8" Feb 14 12:31:53 crc kubenswrapper[4904]: I0214 12:31:53.994907 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-whwk6/crc-debug-9qj7h" Feb 14 12:31:54 crc kubenswrapper[4904]: I0214 12:31:54.326939 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-whwk6/crc-debug-l6rvk"] Feb 14 12:31:54 crc kubenswrapper[4904]: E0214 12:31:54.327296 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43bb26db-1573-4324-bceb-a21e1ab4ceac" containerName="container-00" Feb 14 12:31:54 crc kubenswrapper[4904]: I0214 12:31:54.327310 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="43bb26db-1573-4324-bceb-a21e1ab4ceac" containerName="container-00" Feb 14 12:31:54 crc kubenswrapper[4904]: I0214 12:31:54.327516 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="43bb26db-1573-4324-bceb-a21e1ab4ceac" containerName="container-00" Feb 14 12:31:54 crc kubenswrapper[4904]: I0214 12:31:54.328101 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-whwk6/crc-debug-l6rvk" Feb 14 12:31:54 crc kubenswrapper[4904]: I0214 12:31:54.448130 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wvxf\" (UniqueName: \"kubernetes.io/projected/3d9a3947-562a-4e6b-83d2-2c04e1813641-kube-api-access-9wvxf\") pod \"crc-debug-l6rvk\" (UID: \"3d9a3947-562a-4e6b-83d2-2c04e1813641\") " pod="openshift-must-gather-whwk6/crc-debug-l6rvk" Feb 14 12:31:54 crc kubenswrapper[4904]: I0214 12:31:54.448271 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3d9a3947-562a-4e6b-83d2-2c04e1813641-host\") pod \"crc-debug-l6rvk\" (UID: \"3d9a3947-562a-4e6b-83d2-2c04e1813641\") " pod="openshift-must-gather-whwk6/crc-debug-l6rvk" Feb 14 12:31:54 crc kubenswrapper[4904]: I0214 12:31:54.550022 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wvxf\" (UniqueName: \"kubernetes.io/projected/3d9a3947-562a-4e6b-83d2-2c04e1813641-kube-api-access-9wvxf\") pod \"crc-debug-l6rvk\" (UID: \"3d9a3947-562a-4e6b-83d2-2c04e1813641\") " pod="openshift-must-gather-whwk6/crc-debug-l6rvk" Feb 14 12:31:54 crc kubenswrapper[4904]: I0214 12:31:54.550134 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3d9a3947-562a-4e6b-83d2-2c04e1813641-host\") pod \"crc-debug-l6rvk\" (UID: \"3d9a3947-562a-4e6b-83d2-2c04e1813641\") " pod="openshift-must-gather-whwk6/crc-debug-l6rvk" Feb 14 12:31:54 crc kubenswrapper[4904]: I0214 12:31:54.550294 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3d9a3947-562a-4e6b-83d2-2c04e1813641-host\") pod \"crc-debug-l6rvk\" (UID: \"3d9a3947-562a-4e6b-83d2-2c04e1813641\") " pod="openshift-must-gather-whwk6/crc-debug-l6rvk" Feb 14 12:31:54 crc kubenswrapper[4904]: I0214 12:31:54.569532 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wvxf\" (UniqueName: \"kubernetes.io/projected/3d9a3947-562a-4e6b-83d2-2c04e1813641-kube-api-access-9wvxf\") pod \"crc-debug-l6rvk\" (UID: \"3d9a3947-562a-4e6b-83d2-2c04e1813641\") " pod="openshift-must-gather-whwk6/crc-debug-l6rvk" Feb 14 12:31:54 crc kubenswrapper[4904]: I0214 12:31:54.642709 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-whwk6/crc-debug-l6rvk" Feb 14 12:31:55 crc kubenswrapper[4904]: I0214 12:31:55.003414 4904 generic.go:334] "Generic (PLEG): container finished" podID="3d9a3947-562a-4e6b-83d2-2c04e1813641" containerID="f055d8e837c771bc054a69abfe5a89fd785d8bdc56e484b1e46943ef5a1f5fdf" exitCode=1 Feb 14 12:31:55 crc kubenswrapper[4904]: I0214 12:31:55.003491 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-whwk6/crc-debug-l6rvk" event={"ID":"3d9a3947-562a-4e6b-83d2-2c04e1813641","Type":"ContainerDied","Data":"f055d8e837c771bc054a69abfe5a89fd785d8bdc56e484b1e46943ef5a1f5fdf"} Feb 14 12:31:55 crc kubenswrapper[4904]: I0214 12:31:55.003814 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-whwk6/crc-debug-l6rvk" event={"ID":"3d9a3947-562a-4e6b-83d2-2c04e1813641","Type":"ContainerStarted","Data":"e0ef852cb3ae5ffed0346bb7f7af691584b17dbe425beec1eed2577fbf8d5415"} Feb 14 12:31:55 crc kubenswrapper[4904]: I0214 12:31:55.042578 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-whwk6/crc-debug-l6rvk"] Feb 14 12:31:55 crc kubenswrapper[4904]: I0214 12:31:55.056326 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-whwk6/crc-debug-l6rvk"] Feb 14 12:31:56 crc kubenswrapper[4904]: I0214 12:31:56.103715 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-whwk6/crc-debug-l6rvk" Feb 14 12:31:56 crc kubenswrapper[4904]: I0214 12:31:56.277780 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wvxf\" (UniqueName: \"kubernetes.io/projected/3d9a3947-562a-4e6b-83d2-2c04e1813641-kube-api-access-9wvxf\") pod \"3d9a3947-562a-4e6b-83d2-2c04e1813641\" (UID: \"3d9a3947-562a-4e6b-83d2-2c04e1813641\") " Feb 14 12:31:56 crc kubenswrapper[4904]: I0214 12:31:56.278099 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3d9a3947-562a-4e6b-83d2-2c04e1813641-host\") pod \"3d9a3947-562a-4e6b-83d2-2c04e1813641\" (UID: \"3d9a3947-562a-4e6b-83d2-2c04e1813641\") " Feb 14 12:31:56 crc kubenswrapper[4904]: I0214 12:31:56.278153 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3d9a3947-562a-4e6b-83d2-2c04e1813641-host" (OuterVolumeSpecName: "host") pod "3d9a3947-562a-4e6b-83d2-2c04e1813641" (UID: "3d9a3947-562a-4e6b-83d2-2c04e1813641"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 14 12:31:56 crc kubenswrapper[4904]: I0214 12:31:56.278689 4904 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3d9a3947-562a-4e6b-83d2-2c04e1813641-host\") on node \"crc\" DevicePath \"\"" Feb 14 12:31:56 crc kubenswrapper[4904]: I0214 12:31:56.289089 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d9a3947-562a-4e6b-83d2-2c04e1813641-kube-api-access-9wvxf" (OuterVolumeSpecName: "kube-api-access-9wvxf") pod "3d9a3947-562a-4e6b-83d2-2c04e1813641" (UID: "3d9a3947-562a-4e6b-83d2-2c04e1813641"). InnerVolumeSpecName "kube-api-access-9wvxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:31:56 crc kubenswrapper[4904]: I0214 12:31:56.380929 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wvxf\" (UniqueName: \"kubernetes.io/projected/3d9a3947-562a-4e6b-83d2-2c04e1813641-kube-api-access-9wvxf\") on node \"crc\" DevicePath \"\"" Feb 14 12:31:57 crc kubenswrapper[4904]: I0214 12:31:57.023241 4904 scope.go:117] "RemoveContainer" containerID="f055d8e837c771bc054a69abfe5a89fd785d8bdc56e484b1e46943ef5a1f5fdf" Feb 14 12:31:57 crc kubenswrapper[4904]: I0214 12:31:57.023312 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-whwk6/crc-debug-l6rvk" Feb 14 12:31:57 crc kubenswrapper[4904]: I0214 12:31:57.846736 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d9a3947-562a-4e6b-83d2-2c04e1813641" path="/var/lib/kubelet/pods/3d9a3947-562a-4e6b-83d2-2c04e1813641/volumes" Feb 14 12:32:16 crc kubenswrapper[4904]: I0214 12:32:16.382897 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:32:16 crc kubenswrapper[4904]: I0214 12:32:16.383451 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:32:46 crc kubenswrapper[4904]: I0214 12:32:46.382907 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:32:46 crc kubenswrapper[4904]: I0214 12:32:46.383489 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:33:03 crc kubenswrapper[4904]: I0214 12:33:03.148652 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-c9489b69d-msqgc_fd5faf6b-7dac-4bb3-b496-95d70839ff00/barbican-api/0.log" Feb 14 12:33:03 crc kubenswrapper[4904]: I0214 12:33:03.391366 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6dfd997684-xhxq7_2f74e923-d64a-4161-9839-96a80dd93c2b/barbican-keystone-listener/0.log" Feb 14 12:33:03 crc kubenswrapper[4904]: I0214 12:33:03.419189 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-c9489b69d-msqgc_fd5faf6b-7dac-4bb3-b496-95d70839ff00/barbican-api-log/0.log" Feb 14 12:33:03 crc kubenswrapper[4904]: I0214 12:33:03.467295 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6dfd997684-xhxq7_2f74e923-d64a-4161-9839-96a80dd93c2b/barbican-keystone-listener-log/0.log" Feb 14 12:33:03 crc kubenswrapper[4904]: I0214 12:33:03.612475 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-686dcf6685-s975t_d096f44e-6631-414c-a590-e9e5f2f0cde2/barbican-worker-log/0.log" Feb 14 12:33:03 crc kubenswrapper[4904]: I0214 12:33:03.647760 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-686dcf6685-s975t_d096f44e-6631-414c-a590-e9e5f2f0cde2/barbican-worker/0.log" Feb 14 12:33:03 crc kubenswrapper[4904]: I0214 12:33:03.808920 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_cbe346d4-5e03-40bc-9901-09a4ab6da61c/ceilometer-central-agent/0.log" Feb 14 12:33:04 crc kubenswrapper[4904]: I0214 12:33:04.002132 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_cbe346d4-5e03-40bc-9901-09a4ab6da61c/ceilometer-notification-agent/0.log" Feb 14 12:33:04 crc kubenswrapper[4904]: I0214 12:33:04.033016 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_cbe346d4-5e03-40bc-9901-09a4ab6da61c/sg-core/0.log" Feb 14 12:33:04 crc kubenswrapper[4904]: I0214 12:33:04.056210 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_cbe346d4-5e03-40bc-9901-09a4ab6da61c/proxy-httpd/0.log" Feb 14 12:33:04 crc kubenswrapper[4904]: I0214 12:33:04.271744 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_dd766f2e-4ddd-474b-862e-e0e568132e9c/cinder-api/0.log" Feb 14 12:33:04 crc kubenswrapper[4904]: I0214 12:33:04.293695 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_dd766f2e-4ddd-474b-862e-e0e568132e9c/cinder-api-log/0.log" Feb 14 12:33:04 crc kubenswrapper[4904]: I0214 12:33:04.479210 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_b66062f0-a15b-41cb-b68f-753186581ae2/cinder-scheduler/0.log" Feb 14 12:33:04 crc kubenswrapper[4904]: I0214 12:33:04.573477 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_b66062f0-a15b-41cb-b68f-753186581ae2/probe/0.log" Feb 14 12:33:04 crc kubenswrapper[4904]: I0214 12:33:04.593847 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d7677974f-mz7hn_fa3f3a5c-2d8d-424c-a590-e101c89443a4/init/0.log" Feb 14 12:33:04 crc kubenswrapper[4904]: I0214 12:33:04.834896 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d7677974f-mz7hn_fa3f3a5c-2d8d-424c-a590-e101c89443a4/init/0.log" Feb 14 12:33:04 crc kubenswrapper[4904]: I0214 12:33:04.883541 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d7677974f-mz7hn_fa3f3a5c-2d8d-424c-a590-e101c89443a4/dnsmasq-dns/0.log" Feb 14 12:33:04 crc kubenswrapper[4904]: I0214 12:33:04.928449 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_dba92113-8ae6-46ee-b6ca-32da151c96b0/glance-httpd/0.log" Feb 14 12:33:05 crc kubenswrapper[4904]: I0214 12:33:05.098731 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_dba92113-8ae6-46ee-b6ca-32da151c96b0/glance-log/0.log" Feb 14 12:33:05 crc kubenswrapper[4904]: I0214 12:33:05.149702 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_87c29c32-71ef-48c9-8e26-428071c13df2/glance-httpd/0.log" Feb 14 12:33:05 crc kubenswrapper[4904]: I0214 12:33:05.242424 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_87c29c32-71ef-48c9-8e26-428071c13df2/glance-log/0.log" Feb 14 12:33:05 crc kubenswrapper[4904]: I0214 12:33:05.511231 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6bc9d549b8-hdn6n_cfb252c0-ffb2-4d87-9c4e-b56cc2070d71/horizon/1.log" Feb 14 12:33:05 crc kubenswrapper[4904]: I0214 12:33:05.531218 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6bc9d549b8-hdn6n_cfb252c0-ffb2-4d87-9c4e-b56cc2070d71/horizon/0.log" Feb 14 12:33:05 crc kubenswrapper[4904]: I0214 12:33:05.942140 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7b79cd46c9-g4bx8_ba1e5d40-479d-41ed-953c-3a14733e0673/keystone-api/0.log" Feb 14 12:33:05 crc kubenswrapper[4904]: I0214 12:33:05.956552 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6bc9d549b8-hdn6n_cfb252c0-ffb2-4d87-9c4e-b56cc2070d71/horizon-log/0.log" Feb 14 12:33:06 crc kubenswrapper[4904]: I0214 12:33:06.626470 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29517841-h7rq2_629caf51-0db7-4bd8-80c2-59f45c9e09d2/keystone-cron/0.log" Feb 14 12:33:06 crc kubenswrapper[4904]: I0214 12:33:06.662069 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_e5bdd6b9-9f22-4613-8536-99fdd632d2dd/kube-state-metrics/0.log" Feb 14 12:33:07 crc kubenswrapper[4904]: I0214 12:33:07.031010 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7856f6fd6f-4rhbv_8261bc0e-8af3-4000-aa2d-de7153a235ca/neutron-api/0.log" Feb 14 12:33:07 crc kubenswrapper[4904]: I0214 12:33:07.146339 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7856f6fd6f-4rhbv_8261bc0e-8af3-4000-aa2d-de7153a235ca/neutron-httpd/0.log" Feb 14 12:33:07 crc kubenswrapper[4904]: I0214 12:33:07.726374 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_fe2df707-f559-4976-bda3-a0c2b58cb0cf/nova-api-log/0.log" Feb 14 12:33:07 crc kubenswrapper[4904]: I0214 12:33:07.942307 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_fe2df707-f559-4976-bda3-a0c2b58cb0cf/nova-api-api/0.log" Feb 14 12:33:08 crc kubenswrapper[4904]: I0214 12:33:08.621480 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_1539e687-1fb9-4f6b-ba62-204f09bb691f/nova-cell0-conductor-conductor/0.log" Feb 14 12:33:08 crc kubenswrapper[4904]: I0214 12:33:08.632029 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_fb230e23-e829-4d14-a0ae-9ed889c6840c/nova-cell1-conductor-conductor/0.log" Feb 14 12:33:08 crc kubenswrapper[4904]: I0214 12:33:08.964982 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_aba968e8-e711-4cbc-aea2-6e68c7c9a27c/nova-metadata-log/0.log" Feb 14 12:33:09 crc kubenswrapper[4904]: I0214 12:33:09.033936 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_c89ca73a-9933-4fc8-a96e-48489f786747/nova-cell1-novncproxy-novncproxy/0.log" Feb 14 12:33:09 crc kubenswrapper[4904]: I0214 12:33:09.415950 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_dbbd8d6d-2c1b-49b3-ac1d-6aa39c6994c5/nova-scheduler-scheduler/0.log" Feb 14 12:33:09 crc kubenswrapper[4904]: I0214 12:33:09.531280 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c2878c0a-2515-4412-9aa7-acdb199bc463/mysql-bootstrap/0.log" Feb 14 12:33:09 crc kubenswrapper[4904]: I0214 12:33:09.705487 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c2878c0a-2515-4412-9aa7-acdb199bc463/mysql-bootstrap/0.log" Feb 14 12:33:09 crc kubenswrapper[4904]: I0214 12:33:09.720586 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c2878c0a-2515-4412-9aa7-acdb199bc463/galera/0.log" Feb 14 12:33:10 crc kubenswrapper[4904]: I0214 12:33:10.004897 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b6cb7af3-751c-48dd-84ae-6b77f7b9c233/mysql-bootstrap/0.log" Feb 14 12:33:10 crc kubenswrapper[4904]: I0214 12:33:10.213340 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b6cb7af3-751c-48dd-84ae-6b77f7b9c233/galera/0.log" Feb 14 12:33:10 crc kubenswrapper[4904]: I0214 12:33:10.222879 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b6cb7af3-751c-48dd-84ae-6b77f7b9c233/mysql-bootstrap/0.log" Feb 14 12:33:10 crc kubenswrapper[4904]: I0214 12:33:10.770196 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_ab9aab3c-a75c-4f0a-9125-623bb22314d0/openstackclient/0.log" Feb 14 12:33:10 crc kubenswrapper[4904]: I0214 12:33:10.841931 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-s87wf_692d42dd-2dae-4d7f-bd38-08720eb781bc/openstack-network-exporter/0.log" Feb 14 12:33:10 crc kubenswrapper[4904]: I0214 12:33:10.889214 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_aba968e8-e711-4cbc-aea2-6e68c7c9a27c/nova-metadata-metadata/0.log" Feb 14 12:33:11 crc kubenswrapper[4904]: I0214 12:33:11.064982 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9jk9m_0422124f-385d-4e36-9a12-29bd9a6fa72e/ovsdb-server-init/0.log" Feb 14 12:33:11 crc kubenswrapper[4904]: I0214 12:33:11.289517 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9jk9m_0422124f-385d-4e36-9a12-29bd9a6fa72e/ovsdb-server-init/0.log" Feb 14 12:33:11 crc kubenswrapper[4904]: I0214 12:33:11.381206 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9jk9m_0422124f-385d-4e36-9a12-29bd9a6fa72e/ovs-vswitchd/0.log" Feb 14 12:33:11 crc kubenswrapper[4904]: I0214 12:33:11.462245 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9jk9m_0422124f-385d-4e36-9a12-29bd9a6fa72e/ovsdb-server/0.log" Feb 14 12:33:11 crc kubenswrapper[4904]: I0214 12:33:11.686691 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-zj52b_05a401e2-caea-48a9-bb53-6a42b59dad57/ovn-controller/0.log" Feb 14 12:33:11 crc kubenswrapper[4904]: I0214 12:33:11.754048 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_eed3ea68-257f-49bd-8a2b-4f86a34590c9/ovn-northd/0.log" Feb 14 12:33:11 crc kubenswrapper[4904]: I0214 12:33:11.762987 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_eed3ea68-257f-49bd-8a2b-4f86a34590c9/openstack-network-exporter/0.log" Feb 14 12:33:11 crc kubenswrapper[4904]: I0214 12:33:11.990080 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_02df4ffa-81a6-4fd0-9136-3eedd479f653/ovsdbserver-nb/0.log" Feb 14 12:33:12 crc kubenswrapper[4904]: I0214 12:33:12.009367 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_02df4ffa-81a6-4fd0-9136-3eedd479f653/openstack-network-exporter/0.log" Feb 14 12:33:12 crc kubenswrapper[4904]: I0214 12:33:12.330691 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c/openstack-network-exporter/0.log" Feb 14 12:33:12 crc kubenswrapper[4904]: I0214 12:33:12.395967 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-bb76f8864-hs52l_2dbd404d-9e85-4949-b9a3-194a2708ef89/placement-api/0.log" Feb 14 12:33:12 crc kubenswrapper[4904]: I0214 12:33:12.398553 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_fb07cea3-a6b2-45a1-8aba-fc64dbfa5c5c/ovsdbserver-sb/0.log" Feb 14 12:33:12 crc kubenswrapper[4904]: I0214 12:33:12.674706 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-bb76f8864-hs52l_2dbd404d-9e85-4949-b9a3-194a2708ef89/placement-log/0.log" Feb 14 12:33:12 crc kubenswrapper[4904]: I0214 12:33:12.705796 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_82a277e1-cec5-43a6-a7d4-6651965c02a7/setup-container/0.log" Feb 14 12:33:12 crc kubenswrapper[4904]: I0214 12:33:12.835255 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_82a277e1-cec5-43a6-a7d4-6651965c02a7/setup-container/0.log" Feb 14 12:33:12 crc kubenswrapper[4904]: I0214 12:33:12.987957 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_82a277e1-cec5-43a6-a7d4-6651965c02a7/rabbitmq/0.log" Feb 14 12:33:13 crc kubenswrapper[4904]: I0214 12:33:13.018492 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_87e6f193-7618-42af-a3b3-107cbda62c46/setup-container/0.log" Feb 14 12:33:13 crc kubenswrapper[4904]: I0214 12:33:13.397100 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-49hpl_9243ea89-2a6d-43a8-85d7-b83b3e664680/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 14 12:33:13 crc kubenswrapper[4904]: I0214 12:33:13.398982 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_87e6f193-7618-42af-a3b3-107cbda62c46/rabbitmq/0.log" Feb 14 12:33:13 crc kubenswrapper[4904]: I0214 12:33:13.404544 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_87e6f193-7618-42af-a3b3-107cbda62c46/setup-container/0.log" Feb 14 12:33:13 crc kubenswrapper[4904]: I0214 12:33:13.764494 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-7r6gj_80fcf5ac-b576-4551-8478-d411430ad030/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 14 12:33:13 crc kubenswrapper[4904]: I0214 12:33:13.848300 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-cxfrj_b7e2ea45-3c30-4a1b-b31e-1d4ad38abca6/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 14 12:33:14 crc kubenswrapper[4904]: I0214 12:33:14.017787 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-g8szr_c80fc7fa-a761-4701-8ad2-1843b4421958/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 14 12:33:14 crc kubenswrapper[4904]: I0214 12:33:14.232526 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-gzm2z_245410d0-e53c-4d88-8c47-369c93f449c5/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 14 12:33:14 crc kubenswrapper[4904]: I0214 12:33:14.738182 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-h8gcg_d840dda5-8ddf-4efb-b121-e3113f0302ff/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 14 12:33:14 crc kubenswrapper[4904]: I0214 12:33:14.779865 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-ztxf7_6ec4cf47-ea9f-4e1a-b163-ce7f5c45f4cb/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 14 12:33:15 crc kubenswrapper[4904]: I0214 12:33:15.501269 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5c6f7968ff-b5wb6_b30464d0-5529-441b-8432-0836c20635ea/proxy-server/0.log" Feb 14 12:33:15 crc kubenswrapper[4904]: I0214 12:33:15.690796 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5c6f7968ff-b5wb6_b30464d0-5529-441b-8432-0836c20635ea/proxy-httpd/0.log" Feb 14 12:33:15 crc kubenswrapper[4904]: I0214 12:33:15.782194 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-tw4ls_e2a79513-6beb-47b8-96e5-b5424056661e/swift-ring-rebalance/0.log" Feb 14 12:33:15 crc kubenswrapper[4904]: I0214 12:33:15.967402 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0594337d-464f-489f-9761-3d3d2e143af8/account-auditor/0.log" Feb 14 12:33:15 crc kubenswrapper[4904]: I0214 12:33:15.981635 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0594337d-464f-489f-9761-3d3d2e143af8/account-reaper/0.log" Feb 14 12:33:16 crc kubenswrapper[4904]: I0214 12:33:16.205770 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0594337d-464f-489f-9761-3d3d2e143af8/account-replicator/0.log" Feb 14 12:33:16 crc kubenswrapper[4904]: I0214 12:33:16.225600 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0594337d-464f-489f-9761-3d3d2e143af8/container-auditor/0.log" Feb 14 12:33:16 crc kubenswrapper[4904]: I0214 12:33:16.228026 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0594337d-464f-489f-9761-3d3d2e143af8/account-server/0.log" Feb 14 12:33:16 crc kubenswrapper[4904]: I0214 12:33:16.342158 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0594337d-464f-489f-9761-3d3d2e143af8/container-replicator/0.log" Feb 14 12:33:16 crc kubenswrapper[4904]: I0214 12:33:16.382591 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:33:16 crc kubenswrapper[4904]: I0214 12:33:16.382662 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:33:16 crc kubenswrapper[4904]: I0214 12:33:16.382711 4904 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 12:33:16 crc kubenswrapper[4904]: I0214 12:33:16.383561 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658"} pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 14 12:33:16 crc kubenswrapper[4904]: I0214 12:33:16.383655 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" containerID="cri-o://ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" gracePeriod=600 Feb 14 12:33:16 crc kubenswrapper[4904]: E0214 12:33:16.725484 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:33:16 crc kubenswrapper[4904]: E0214 12:33:16.917941 4904 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7790d276_2506_4641_8c22_67e6441417f0.slice/crio-conmon-ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658.scope\": RecentStats: unable to find data in memory cache]" Feb 14 12:33:16 crc kubenswrapper[4904]: I0214 12:33:16.931214 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0594337d-464f-489f-9761-3d3d2e143af8/object-expirer/0.log" Feb 14 12:33:16 crc kubenswrapper[4904]: I0214 12:33:16.953469 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0594337d-464f-489f-9761-3d3d2e143af8/container-server/0.log" Feb 14 12:33:16 crc kubenswrapper[4904]: I0214 12:33:16.991689 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0594337d-464f-489f-9761-3d3d2e143af8/container-updater/0.log" Feb 14 12:33:17 crc kubenswrapper[4904]: I0214 12:33:17.019880 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0594337d-464f-489f-9761-3d3d2e143af8/object-auditor/0.log" Feb 14 12:33:17 crc kubenswrapper[4904]: I0214 12:33:17.219090 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0594337d-464f-489f-9761-3d3d2e143af8/object-replicator/0.log" Feb 14 12:33:17 crc kubenswrapper[4904]: I0214 12:33:17.219227 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0594337d-464f-489f-9761-3d3d2e143af8/object-server/0.log" Feb 14 12:33:17 crc kubenswrapper[4904]: I0214 12:33:17.284246 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0594337d-464f-489f-9761-3d3d2e143af8/object-updater/0.log" Feb 14 12:33:17 crc kubenswrapper[4904]: I0214 12:33:17.331570 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0594337d-464f-489f-9761-3d3d2e143af8/rsync/0.log" Feb 14 12:33:17 crc kubenswrapper[4904]: I0214 12:33:17.442564 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0594337d-464f-489f-9761-3d3d2e143af8/swift-recon-cron/0.log" Feb 14 12:33:17 crc kubenswrapper[4904]: I0214 12:33:17.682017 4904 generic.go:334] "Generic (PLEG): container finished" podID="7790d276-2506-4641-8c22-67e6441417f0" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" exitCode=0 Feb 14 12:33:17 crc kubenswrapper[4904]: I0214 12:33:17.682066 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerDied","Data":"ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658"} Feb 14 12:33:17 crc kubenswrapper[4904]: I0214 12:33:17.682101 4904 scope.go:117] "RemoveContainer" containerID="c79f60f3a3c1cb61784ba260135e84fa21d0779eb625031dd8470a43f1ee7ee4" Feb 14 12:33:17 crc kubenswrapper[4904]: I0214 12:33:17.682747 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:33:17 crc kubenswrapper[4904]: E0214 12:33:17.683091 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:33:20 crc kubenswrapper[4904]: I0214 12:33:20.261934 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_a35e40d9-4ced-4411-aff4-68c295331d47/memcached/0.log" Feb 14 12:33:28 crc kubenswrapper[4904]: I0214 12:33:28.836546 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:33:28 crc kubenswrapper[4904]: E0214 12:33:28.837263 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:33:39 crc kubenswrapper[4904]: I0214 12:33:39.841644 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:33:39 crc kubenswrapper[4904]: E0214 12:33:39.842372 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:33:47 crc kubenswrapper[4904]: I0214 12:33:47.869349 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d8bf5c495-gt547_6c57743e-c864-4e0d-828a-406978aab771/manager/0.log" Feb 14 12:33:48 crc kubenswrapper[4904]: I0214 12:33:48.260510 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn_a0bea6e6-38b1-427d-8699-a29ff81030a7/util/0.log" Feb 14 12:33:48 crc kubenswrapper[4904]: I0214 12:33:48.484658 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn_a0bea6e6-38b1-427d-8699-a29ff81030a7/util/0.log" Feb 14 12:33:48 crc kubenswrapper[4904]: I0214 12:33:48.515359 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn_a0bea6e6-38b1-427d-8699-a29ff81030a7/pull/0.log" Feb 14 12:33:49 crc kubenswrapper[4904]: I0214 12:33:49.336883 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn_a0bea6e6-38b1-427d-8699-a29ff81030a7/pull/0.log" Feb 14 12:33:49 crc kubenswrapper[4904]: I0214 12:33:49.554333 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn_a0bea6e6-38b1-427d-8699-a29ff81030a7/util/0.log" Feb 14 12:33:49 crc kubenswrapper[4904]: I0214 12:33:49.849244 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5d946d989d-hbqp9_33670c4d-fd2a-49cd-b602-ccdd927b7c00/manager/0.log" Feb 14 12:33:49 crc kubenswrapper[4904]: I0214 12:33:49.887151 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn_a0bea6e6-38b1-427d-8699-a29ff81030a7/pull/0.log" Feb 14 12:33:50 crc kubenswrapper[4904]: I0214 12:33:50.128872 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e1ed6567609b8d07a4443887b95e953c8e537976bcfb12a9d14468d61cxkhxn_a0bea6e6-38b1-427d-8699-a29ff81030a7/extract/0.log" Feb 14 12:33:50 crc kubenswrapper[4904]: I0214 12:33:50.285207 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987464f4-jsn2d_2db31100-df1f-441c-87bf-944dc247a4fc/manager/0.log" Feb 14 12:33:50 crc kubenswrapper[4904]: I0214 12:33:50.643146 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69f49c598c-mbrvg_96c0bf26-aa73-4948-add8-b30fbbc23df9/manager/0.log" Feb 14 12:33:51 crc kubenswrapper[4904]: I0214 12:33:51.085724 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b9b8895d5-m8p4r_d4632ec1-624e-4097-9888-0bbe395d6916/manager/0.log" Feb 14 12:33:51 crc kubenswrapper[4904]: I0214 12:33:51.544448 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-554564d7fc-k28jp_8510a96b-51e8-4ad4-be1e-0fa7e2a86880/manager/0.log" Feb 14 12:33:51 crc kubenswrapper[4904]: I0214 12:33:51.750794 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5478cb7c68-h92m8_8b9716ac-90dd-48ff-9dc8-7e5ef53ce0e9/manager/0.log" Feb 14 12:33:51 crc kubenswrapper[4904]: I0214 12:33:51.838426 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:33:51 crc kubenswrapper[4904]: E0214 12:33:51.838716 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:33:51 crc kubenswrapper[4904]: I0214 12:33:51.912757 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b4d948c87-sjlh5_56b8547b-7339-452b-9a06-a05411222f70/manager/0.log" Feb 14 12:33:52 crc kubenswrapper[4904]: I0214 12:33:52.056555 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-54f6768c69-6hjt5_23733893-aa4e-47e2-a06a-a055313308d8/manager/0.log" Feb 14 12:33:52 crc kubenswrapper[4904]: I0214 12:33:52.438364 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6994f66f48-rvcg7_b80a5769-6d09-4697-ba97-a5723e161a3c/manager/0.log" Feb 14 12:33:52 crc kubenswrapper[4904]: I0214 12:33:52.636226 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64ddbf8bb-hsx4c_bac77172-b950-473c-8b07-0b1201a981db/manager/0.log" Feb 14 12:33:52 crc kubenswrapper[4904]: I0214 12:33:52.873715 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-567668f5cf-qd78d_f29b166d-edb4-4403-be20-a523578848ee/manager/0.log" Feb 14 12:33:53 crc kubenswrapper[4904]: I0214 12:33:53.030366 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84966cf5c46s685_c3d1040e-0300-49ce-8b7d-7e4dac11c874/manager/0.log" Feb 14 12:33:53 crc kubenswrapper[4904]: I0214 12:33:53.573949 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-bfd897969-pqr4v_6544b724-ca9a-4dec-9d1e-987e8b9a375b/operator/0.log" Feb 14 12:33:54 crc kubenswrapper[4904]: I0214 12:33:54.320800 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-r9h6q_b429a7c8-df33-430d-9d4f-a678e2639575/registry-server/0.log" Feb 14 12:33:55 crc kubenswrapper[4904]: I0214 12:33:55.054058 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-d44cf6b75-9gwm7_733ceff2-fece-49be-8bc4-63f3adcfff22/manager/0.log" Feb 14 12:33:55 crc kubenswrapper[4904]: I0214 12:33:55.455700 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8497b45c89-pwsvp_70a8fe80-89cb-492b-a631-e8042e3ed1be/manager/0.log" Feb 14 12:33:55 crc kubenswrapper[4904]: I0214 12:33:55.696038 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f8888797-xb6dd_cebffdc5-38fd-45a5-ab0a-7445696c1661/manager/0.log" Feb 14 12:33:55 crc kubenswrapper[4904]: I0214 12:33:55.804118 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-795l4_aeb77a69-e9d9-4e04-a270-665a28bb0fa1/operator/0.log" Feb 14 12:33:55 crc kubenswrapper[4904]: I0214 12:33:55.928337 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-67d96975f-hzkn7_850dfcee-db10-493e-8043-5e167cfe3a54/manager/0.log" Feb 14 12:33:56 crc kubenswrapper[4904]: I0214 12:33:56.063179 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68f46476f-4zj5b_fa0069c3-fd8a-4ffe-aad0-ccee78f70caa/manager/0.log" Feb 14 12:33:56 crc kubenswrapper[4904]: I0214 12:33:56.266139 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f45b4ff68-q5cmd_95a68c48-4e90-43c0-9caf-d6be291fca1e/manager/0.log" Feb 14 12:33:56 crc kubenswrapper[4904]: I0214 12:33:56.394785 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7866795846-jnd94_00d1bd7d-c3d8-4a54-9208-c28355796751/manager/0.log" Feb 14 12:33:56 crc kubenswrapper[4904]: I0214 12:33:56.564670 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5db88f68c-x8k6m_53ba6432-a3b1-4cce-8ab2-bc74a718b134/manager/0.log" Feb 14 12:33:58 crc kubenswrapper[4904]: I0214 12:33:58.581402 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-868647ff47-m29fl_2b26d95c-682d-4be5-9788-5044e186611c/manager/0.log" Feb 14 12:34:04 crc kubenswrapper[4904]: I0214 12:34:04.836929 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:34:04 crc kubenswrapper[4904]: E0214 12:34:04.837613 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:34:06 crc kubenswrapper[4904]: I0214 12:34:06.529776 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tpv7v"] Feb 14 12:34:06 crc kubenswrapper[4904]: E0214 12:34:06.530219 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d9a3947-562a-4e6b-83d2-2c04e1813641" containerName="container-00" Feb 14 12:34:06 crc kubenswrapper[4904]: I0214 12:34:06.530235 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d9a3947-562a-4e6b-83d2-2c04e1813641" containerName="container-00" Feb 14 12:34:06 crc kubenswrapper[4904]: I0214 12:34:06.530448 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d9a3947-562a-4e6b-83d2-2c04e1813641" containerName="container-00" Feb 14 12:34:06 crc kubenswrapper[4904]: I0214 12:34:06.538058 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tpv7v" Feb 14 12:34:06 crc kubenswrapper[4904]: I0214 12:34:06.571896 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tpv7v"] Feb 14 12:34:06 crc kubenswrapper[4904]: I0214 12:34:06.637157 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l58kq\" (UniqueName: \"kubernetes.io/projected/ad6c70dd-1fb9-416c-aebf-15af183a1b4c-kube-api-access-l58kq\") pod \"community-operators-tpv7v\" (UID: \"ad6c70dd-1fb9-416c-aebf-15af183a1b4c\") " pod="openshift-marketplace/community-operators-tpv7v" Feb 14 12:34:06 crc kubenswrapper[4904]: I0214 12:34:06.637493 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad6c70dd-1fb9-416c-aebf-15af183a1b4c-utilities\") pod \"community-operators-tpv7v\" (UID: \"ad6c70dd-1fb9-416c-aebf-15af183a1b4c\") " pod="openshift-marketplace/community-operators-tpv7v" Feb 14 12:34:06 crc kubenswrapper[4904]: I0214 12:34:06.637608 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad6c70dd-1fb9-416c-aebf-15af183a1b4c-catalog-content\") pod \"community-operators-tpv7v\" (UID: \"ad6c70dd-1fb9-416c-aebf-15af183a1b4c\") " pod="openshift-marketplace/community-operators-tpv7v" Feb 14 12:34:06 crc kubenswrapper[4904]: I0214 12:34:06.739725 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad6c70dd-1fb9-416c-aebf-15af183a1b4c-utilities\") pod \"community-operators-tpv7v\" (UID: \"ad6c70dd-1fb9-416c-aebf-15af183a1b4c\") " pod="openshift-marketplace/community-operators-tpv7v" Feb 14 12:34:06 crc kubenswrapper[4904]: I0214 12:34:06.740024 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad6c70dd-1fb9-416c-aebf-15af183a1b4c-catalog-content\") pod \"community-operators-tpv7v\" (UID: \"ad6c70dd-1fb9-416c-aebf-15af183a1b4c\") " pod="openshift-marketplace/community-operators-tpv7v" Feb 14 12:34:06 crc kubenswrapper[4904]: I0214 12:34:06.740154 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l58kq\" (UniqueName: \"kubernetes.io/projected/ad6c70dd-1fb9-416c-aebf-15af183a1b4c-kube-api-access-l58kq\") pod \"community-operators-tpv7v\" (UID: \"ad6c70dd-1fb9-416c-aebf-15af183a1b4c\") " pod="openshift-marketplace/community-operators-tpv7v" Feb 14 12:34:06 crc kubenswrapper[4904]: I0214 12:34:06.740249 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad6c70dd-1fb9-416c-aebf-15af183a1b4c-utilities\") pod \"community-operators-tpv7v\" (UID: \"ad6c70dd-1fb9-416c-aebf-15af183a1b4c\") " pod="openshift-marketplace/community-operators-tpv7v" Feb 14 12:34:06 crc kubenswrapper[4904]: I0214 12:34:06.740278 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad6c70dd-1fb9-416c-aebf-15af183a1b4c-catalog-content\") pod \"community-operators-tpv7v\" (UID: \"ad6c70dd-1fb9-416c-aebf-15af183a1b4c\") " pod="openshift-marketplace/community-operators-tpv7v" Feb 14 12:34:06 crc kubenswrapper[4904]: I0214 12:34:06.762661 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l58kq\" (UniqueName: \"kubernetes.io/projected/ad6c70dd-1fb9-416c-aebf-15af183a1b4c-kube-api-access-l58kq\") pod \"community-operators-tpv7v\" (UID: \"ad6c70dd-1fb9-416c-aebf-15af183a1b4c\") " pod="openshift-marketplace/community-operators-tpv7v" Feb 14 12:34:06 crc kubenswrapper[4904]: I0214 12:34:06.868540 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tpv7v" Feb 14 12:34:07 crc kubenswrapper[4904]: I0214 12:34:07.466557 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tpv7v"] Feb 14 12:34:08 crc kubenswrapper[4904]: I0214 12:34:08.130040 4904 generic.go:334] "Generic (PLEG): container finished" podID="ad6c70dd-1fb9-416c-aebf-15af183a1b4c" containerID="2d4ae63aceaf9d9616c447e54153c3145f529888a653346a622e63dbc2a391fa" exitCode=0 Feb 14 12:34:08 crc kubenswrapper[4904]: I0214 12:34:08.130099 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpv7v" event={"ID":"ad6c70dd-1fb9-416c-aebf-15af183a1b4c","Type":"ContainerDied","Data":"2d4ae63aceaf9d9616c447e54153c3145f529888a653346a622e63dbc2a391fa"} Feb 14 12:34:08 crc kubenswrapper[4904]: I0214 12:34:08.130335 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpv7v" event={"ID":"ad6c70dd-1fb9-416c-aebf-15af183a1b4c","Type":"ContainerStarted","Data":"50a424b1269710bf90205eee4d466d4b91cae6dd3b882e176ddf9a0957eee491"} Feb 14 12:34:09 crc kubenswrapper[4904]: I0214 12:34:09.139520 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpv7v" event={"ID":"ad6c70dd-1fb9-416c-aebf-15af183a1b4c","Type":"ContainerStarted","Data":"de33587daad6e6506a03d6244da829ce500056e307335b411bb1af7aa64d57fe"} Feb 14 12:34:11 crc kubenswrapper[4904]: I0214 12:34:11.156417 4904 generic.go:334] "Generic (PLEG): container finished" podID="ad6c70dd-1fb9-416c-aebf-15af183a1b4c" containerID="de33587daad6e6506a03d6244da829ce500056e307335b411bb1af7aa64d57fe" exitCode=0 Feb 14 12:34:11 crc kubenswrapper[4904]: I0214 12:34:11.156501 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpv7v" event={"ID":"ad6c70dd-1fb9-416c-aebf-15af183a1b4c","Type":"ContainerDied","Data":"de33587daad6e6506a03d6244da829ce500056e307335b411bb1af7aa64d57fe"} Feb 14 12:34:12 crc kubenswrapper[4904]: I0214 12:34:12.167220 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpv7v" event={"ID":"ad6c70dd-1fb9-416c-aebf-15af183a1b4c","Type":"ContainerStarted","Data":"b7891a4f47591be68fa2665a8ad273d687697933c8a357e2011123fc814f3b01"} Feb 14 12:34:12 crc kubenswrapper[4904]: I0214 12:34:12.189937 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tpv7v" podStartSLOduration=2.759713239 podStartE2EDuration="6.189919564s" podCreationTimestamp="2026-02-14 12:34:06 +0000 UTC" firstStartedPulling="2026-02-14 12:34:08.137079723 +0000 UTC m=+5038.949844384" lastFinishedPulling="2026-02-14 12:34:11.567286048 +0000 UTC m=+5042.380050709" observedRunningTime="2026-02-14 12:34:12.18389772 +0000 UTC m=+5042.996662381" watchObservedRunningTime="2026-02-14 12:34:12.189919564 +0000 UTC m=+5043.002684225" Feb 14 12:34:16 crc kubenswrapper[4904]: I0214 12:34:16.836349 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:34:16 crc kubenswrapper[4904]: E0214 12:34:16.837053 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:34:16 crc kubenswrapper[4904]: I0214 12:34:16.869721 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tpv7v" Feb 14 12:34:16 crc kubenswrapper[4904]: I0214 12:34:16.870191 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tpv7v" Feb 14 12:34:17 crc kubenswrapper[4904]: I0214 12:34:17.928868 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-tpv7v" podUID="ad6c70dd-1fb9-416c-aebf-15af183a1b4c" containerName="registry-server" probeResult="failure" output=< Feb 14 12:34:17 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 12:34:17 crc kubenswrapper[4904]: > Feb 14 12:34:20 crc kubenswrapper[4904]: I0214 12:34:20.441623 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-sjcqb_ca75271d-7ae7-4a71-af10-732866e33d74/control-plane-machine-set-operator/0.log" Feb 14 12:34:20 crc kubenswrapper[4904]: I0214 12:34:20.707190 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-hgggh_800acf17-404a-4dc3-a25f-66e779d292fd/kube-rbac-proxy/0.log" Feb 14 12:34:20 crc kubenswrapper[4904]: I0214 12:34:20.707765 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-hgggh_800acf17-404a-4dc3-a25f-66e779d292fd/machine-api-operator/0.log" Feb 14 12:34:26 crc kubenswrapper[4904]: I0214 12:34:26.929070 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tpv7v" Feb 14 12:34:26 crc kubenswrapper[4904]: I0214 12:34:26.986111 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tpv7v" Feb 14 12:34:28 crc kubenswrapper[4904]: I0214 12:34:28.009195 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tpv7v"] Feb 14 12:34:28 crc kubenswrapper[4904]: I0214 12:34:28.304980 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tpv7v" podUID="ad6c70dd-1fb9-416c-aebf-15af183a1b4c" containerName="registry-server" containerID="cri-o://b7891a4f47591be68fa2665a8ad273d687697933c8a357e2011123fc814f3b01" gracePeriod=2 Feb 14 12:34:28 crc kubenswrapper[4904]: I0214 12:34:28.837721 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:34:28 crc kubenswrapper[4904]: E0214 12:34:28.838336 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:34:28 crc kubenswrapper[4904]: I0214 12:34:28.946953 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tpv7v" Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.073951 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad6c70dd-1fb9-416c-aebf-15af183a1b4c-catalog-content\") pod \"ad6c70dd-1fb9-416c-aebf-15af183a1b4c\" (UID: \"ad6c70dd-1fb9-416c-aebf-15af183a1b4c\") " Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.074163 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad6c70dd-1fb9-416c-aebf-15af183a1b4c-utilities\") pod \"ad6c70dd-1fb9-416c-aebf-15af183a1b4c\" (UID: \"ad6c70dd-1fb9-416c-aebf-15af183a1b4c\") " Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.074193 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l58kq\" (UniqueName: \"kubernetes.io/projected/ad6c70dd-1fb9-416c-aebf-15af183a1b4c-kube-api-access-l58kq\") pod \"ad6c70dd-1fb9-416c-aebf-15af183a1b4c\" (UID: \"ad6c70dd-1fb9-416c-aebf-15af183a1b4c\") " Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.074900 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad6c70dd-1fb9-416c-aebf-15af183a1b4c-utilities" (OuterVolumeSpecName: "utilities") pod "ad6c70dd-1fb9-416c-aebf-15af183a1b4c" (UID: "ad6c70dd-1fb9-416c-aebf-15af183a1b4c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.096306 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad6c70dd-1fb9-416c-aebf-15af183a1b4c-kube-api-access-l58kq" (OuterVolumeSpecName: "kube-api-access-l58kq") pod "ad6c70dd-1fb9-416c-aebf-15af183a1b4c" (UID: "ad6c70dd-1fb9-416c-aebf-15af183a1b4c"). InnerVolumeSpecName "kube-api-access-l58kq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.142661 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad6c70dd-1fb9-416c-aebf-15af183a1b4c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad6c70dd-1fb9-416c-aebf-15af183a1b4c" (UID: "ad6c70dd-1fb9-416c-aebf-15af183a1b4c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.176580 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad6c70dd-1fb9-416c-aebf-15af183a1b4c-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.176624 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l58kq\" (UniqueName: \"kubernetes.io/projected/ad6c70dd-1fb9-416c-aebf-15af183a1b4c-kube-api-access-l58kq\") on node \"crc\" DevicePath \"\"" Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.176638 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad6c70dd-1fb9-416c-aebf-15af183a1b4c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.318427 4904 generic.go:334] "Generic (PLEG): container finished" podID="ad6c70dd-1fb9-416c-aebf-15af183a1b4c" containerID="b7891a4f47591be68fa2665a8ad273d687697933c8a357e2011123fc814f3b01" exitCode=0 Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.318483 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpv7v" event={"ID":"ad6c70dd-1fb9-416c-aebf-15af183a1b4c","Type":"ContainerDied","Data":"b7891a4f47591be68fa2665a8ad273d687697933c8a357e2011123fc814f3b01"} Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.318520 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpv7v" event={"ID":"ad6c70dd-1fb9-416c-aebf-15af183a1b4c","Type":"ContainerDied","Data":"50a424b1269710bf90205eee4d466d4b91cae6dd3b882e176ddf9a0957eee491"} Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.318541 4904 scope.go:117] "RemoveContainer" containerID="b7891a4f47591be68fa2665a8ad273d687697933c8a357e2011123fc814f3b01" Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.318739 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tpv7v" Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.359017 4904 scope.go:117] "RemoveContainer" containerID="de33587daad6e6506a03d6244da829ce500056e307335b411bb1af7aa64d57fe" Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.367945 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tpv7v"] Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.374919 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tpv7v"] Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.388460 4904 scope.go:117] "RemoveContainer" containerID="2d4ae63aceaf9d9616c447e54153c3145f529888a653346a622e63dbc2a391fa" Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.441892 4904 scope.go:117] "RemoveContainer" containerID="b7891a4f47591be68fa2665a8ad273d687697933c8a357e2011123fc814f3b01" Feb 14 12:34:29 crc kubenswrapper[4904]: E0214 12:34:29.442541 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7891a4f47591be68fa2665a8ad273d687697933c8a357e2011123fc814f3b01\": container with ID starting with b7891a4f47591be68fa2665a8ad273d687697933c8a357e2011123fc814f3b01 not found: ID does not exist" containerID="b7891a4f47591be68fa2665a8ad273d687697933c8a357e2011123fc814f3b01" Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.442576 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7891a4f47591be68fa2665a8ad273d687697933c8a357e2011123fc814f3b01"} err="failed to get container status \"b7891a4f47591be68fa2665a8ad273d687697933c8a357e2011123fc814f3b01\": rpc error: code = NotFound desc = could not find container \"b7891a4f47591be68fa2665a8ad273d687697933c8a357e2011123fc814f3b01\": container with ID starting with b7891a4f47591be68fa2665a8ad273d687697933c8a357e2011123fc814f3b01 not found: ID does not exist" Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.442596 4904 scope.go:117] "RemoveContainer" containerID="de33587daad6e6506a03d6244da829ce500056e307335b411bb1af7aa64d57fe" Feb 14 12:34:29 crc kubenswrapper[4904]: E0214 12:34:29.443103 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de33587daad6e6506a03d6244da829ce500056e307335b411bb1af7aa64d57fe\": container with ID starting with de33587daad6e6506a03d6244da829ce500056e307335b411bb1af7aa64d57fe not found: ID does not exist" containerID="de33587daad6e6506a03d6244da829ce500056e307335b411bb1af7aa64d57fe" Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.443148 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de33587daad6e6506a03d6244da829ce500056e307335b411bb1af7aa64d57fe"} err="failed to get container status \"de33587daad6e6506a03d6244da829ce500056e307335b411bb1af7aa64d57fe\": rpc error: code = NotFound desc = could not find container \"de33587daad6e6506a03d6244da829ce500056e307335b411bb1af7aa64d57fe\": container with ID starting with de33587daad6e6506a03d6244da829ce500056e307335b411bb1af7aa64d57fe not found: ID does not exist" Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.443177 4904 scope.go:117] "RemoveContainer" containerID="2d4ae63aceaf9d9616c447e54153c3145f529888a653346a622e63dbc2a391fa" Feb 14 12:34:29 crc kubenswrapper[4904]: E0214 12:34:29.443498 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d4ae63aceaf9d9616c447e54153c3145f529888a653346a622e63dbc2a391fa\": container with ID starting with 2d4ae63aceaf9d9616c447e54153c3145f529888a653346a622e63dbc2a391fa not found: ID does not exist" containerID="2d4ae63aceaf9d9616c447e54153c3145f529888a653346a622e63dbc2a391fa" Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.443525 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d4ae63aceaf9d9616c447e54153c3145f529888a653346a622e63dbc2a391fa"} err="failed to get container status \"2d4ae63aceaf9d9616c447e54153c3145f529888a653346a622e63dbc2a391fa\": rpc error: code = NotFound desc = could not find container \"2d4ae63aceaf9d9616c447e54153c3145f529888a653346a622e63dbc2a391fa\": container with ID starting with 2d4ae63aceaf9d9616c447e54153c3145f529888a653346a622e63dbc2a391fa not found: ID does not exist" Feb 14 12:34:29 crc kubenswrapper[4904]: I0214 12:34:29.847075 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad6c70dd-1fb9-416c-aebf-15af183a1b4c" path="/var/lib/kubelet/pods/ad6c70dd-1fb9-416c-aebf-15af183a1b4c/volumes" Feb 14 12:34:35 crc kubenswrapper[4904]: I0214 12:34:35.695198 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-8qkmb_34109571-3f63-4e10-b5ed-ed86c6713f24/cert-manager-controller/0.log" Feb 14 12:34:35 crc kubenswrapper[4904]: I0214 12:34:35.958098 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-zwf2w_b643d341-7020-4a44-9c9a-bd60f9968f90/cert-manager-cainjector/0.log" Feb 14 12:34:36 crc kubenswrapper[4904]: I0214 12:34:36.014759 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-rr2vn_25466585-eaf0-4056-9bd2-766da897ddf6/cert-manager-webhook/0.log" Feb 14 12:34:41 crc kubenswrapper[4904]: I0214 12:34:41.836175 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:34:41 crc kubenswrapper[4904]: E0214 12:34:41.837124 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:34:48 crc kubenswrapper[4904]: I0214 12:34:48.547250 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5c78fc5d65-pc5mk_3d6f625a-0a1c-4bcc-96e6-0e756fd3be6a/nmstate-console-plugin/0.log" Feb 14 12:34:48 crc kubenswrapper[4904]: I0214 12:34:48.697668 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-g5z5r_967e4fe2-b2d0-4526-8d4d-1c0e4c3e6f62/nmstate-handler/0.log" Feb 14 12:34:48 crc kubenswrapper[4904]: I0214 12:34:48.842684 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-js9s2_87d08820-f798-415f-a367-e1ec7bfb86d2/nmstate-metrics/0.log" Feb 14 12:34:48 crc kubenswrapper[4904]: I0214 12:34:48.849195 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-js9s2_87d08820-f798-415f-a367-e1ec7bfb86d2/kube-rbac-proxy/0.log" Feb 14 12:34:49 crc kubenswrapper[4904]: I0214 12:34:49.125433 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-694c9596b7-nxxlw_38f7c921-1f4f-4217-92db-94809df6750c/nmstate-operator/0.log" Feb 14 12:34:49 crc kubenswrapper[4904]: I0214 12:34:49.153918 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-866bcb46dc-tj7n4_62ca493b-5a80-450e-9e45-8f916f2c05e8/nmstate-webhook/0.log" Feb 14 12:34:54 crc kubenswrapper[4904]: I0214 12:34:54.837009 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:34:54 crc kubenswrapper[4904]: E0214 12:34:54.838108 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:35:09 crc kubenswrapper[4904]: I0214 12:35:09.848404 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:35:09 crc kubenswrapper[4904]: E0214 12:35:09.849176 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:35:15 crc kubenswrapper[4904]: I0214 12:35:15.629299 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-bf7zh_ed02a813-b989-405f-87d8-6b4ed3539671/kube-rbac-proxy/0.log" Feb 14 12:35:15 crc kubenswrapper[4904]: I0214 12:35:15.722438 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-bf7zh_ed02a813-b989-405f-87d8-6b4ed3539671/controller/0.log" Feb 14 12:35:15 crc kubenswrapper[4904]: I0214 12:35:15.872710 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-78b44bf5bb-jzq7k_b15d7113-f0f8-4033-a293-fc94960799c4/frr-k8s-webhook-server/0.log" Feb 14 12:35:16 crc kubenswrapper[4904]: I0214 12:35:16.000104 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6sv9_bdf867d3-84cd-4b7c-8e49-97ba408a1689/cp-frr-files/0.log" Feb 14 12:35:16 crc kubenswrapper[4904]: I0214 12:35:16.176212 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6sv9_bdf867d3-84cd-4b7c-8e49-97ba408a1689/cp-frr-files/0.log" Feb 14 12:35:16 crc kubenswrapper[4904]: I0214 12:35:16.206182 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6sv9_bdf867d3-84cd-4b7c-8e49-97ba408a1689/cp-reloader/0.log" Feb 14 12:35:16 crc kubenswrapper[4904]: I0214 12:35:16.234131 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6sv9_bdf867d3-84cd-4b7c-8e49-97ba408a1689/cp-reloader/0.log" Feb 14 12:35:16 crc kubenswrapper[4904]: I0214 12:35:16.262351 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6sv9_bdf867d3-84cd-4b7c-8e49-97ba408a1689/cp-metrics/0.log" Feb 14 12:35:16 crc kubenswrapper[4904]: I0214 12:35:16.407170 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6sv9_bdf867d3-84cd-4b7c-8e49-97ba408a1689/cp-frr-files/0.log" Feb 14 12:35:16 crc kubenswrapper[4904]: I0214 12:35:16.413498 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6sv9_bdf867d3-84cd-4b7c-8e49-97ba408a1689/cp-metrics/0.log" Feb 14 12:35:16 crc kubenswrapper[4904]: I0214 12:35:16.439607 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6sv9_bdf867d3-84cd-4b7c-8e49-97ba408a1689/cp-reloader/0.log" Feb 14 12:35:16 crc kubenswrapper[4904]: I0214 12:35:16.505294 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6sv9_bdf867d3-84cd-4b7c-8e49-97ba408a1689/cp-metrics/0.log" Feb 14 12:35:16 crc kubenswrapper[4904]: I0214 12:35:16.636827 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6sv9_bdf867d3-84cd-4b7c-8e49-97ba408a1689/cp-metrics/0.log" Feb 14 12:35:16 crc kubenswrapper[4904]: I0214 12:35:16.673046 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6sv9_bdf867d3-84cd-4b7c-8e49-97ba408a1689/cp-frr-files/0.log" Feb 14 12:35:16 crc kubenswrapper[4904]: I0214 12:35:16.702718 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6sv9_bdf867d3-84cd-4b7c-8e49-97ba408a1689/cp-reloader/0.log" Feb 14 12:35:16 crc kubenswrapper[4904]: I0214 12:35:16.710942 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6sv9_bdf867d3-84cd-4b7c-8e49-97ba408a1689/controller/0.log" Feb 14 12:35:16 crc kubenswrapper[4904]: I0214 12:35:16.870693 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6sv9_bdf867d3-84cd-4b7c-8e49-97ba408a1689/frr-metrics/0.log" Feb 14 12:35:16 crc kubenswrapper[4904]: I0214 12:35:16.979577 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6sv9_bdf867d3-84cd-4b7c-8e49-97ba408a1689/kube-rbac-proxy/0.log" Feb 14 12:35:17 crc kubenswrapper[4904]: I0214 12:35:17.015193 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6sv9_bdf867d3-84cd-4b7c-8e49-97ba408a1689/kube-rbac-proxy-frr/0.log" Feb 14 12:35:17 crc kubenswrapper[4904]: I0214 12:35:17.212728 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6sv9_bdf867d3-84cd-4b7c-8e49-97ba408a1689/reloader/0.log" Feb 14 12:35:17 crc kubenswrapper[4904]: I0214 12:35:17.239469 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7b865f4cbf-8thhc_678c1429-9ad4-44cd-8f86-29bd34395c5b/manager/0.log" Feb 14 12:35:17 crc kubenswrapper[4904]: I0214 12:35:17.501455 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-d55fb6f5-cljdr_481cda22-3fb4-4c12-a4ed-29b316a69915/webhook-server/0.log" Feb 14 12:35:17 crc kubenswrapper[4904]: I0214 12:35:17.729042 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zmg22_90c810e2-d81e-44b0-bf0c-ec3ba2776c55/kube-rbac-proxy/0.log" Feb 14 12:35:18 crc kubenswrapper[4904]: I0214 12:35:18.238211 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-x6sv9_bdf867d3-84cd-4b7c-8e49-97ba408a1689/frr/0.log" Feb 14 12:35:18 crc kubenswrapper[4904]: I0214 12:35:18.259911 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zmg22_90c810e2-d81e-44b0-bf0c-ec3ba2776c55/speaker/0.log" Feb 14 12:35:20 crc kubenswrapper[4904]: I0214 12:35:20.836973 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:35:20 crc kubenswrapper[4904]: E0214 12:35:20.837308 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:35:28 crc kubenswrapper[4904]: I0214 12:35:28.126052 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-98pb6"] Feb 14 12:35:28 crc kubenswrapper[4904]: E0214 12:35:28.128150 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad6c70dd-1fb9-416c-aebf-15af183a1b4c" containerName="extract-content" Feb 14 12:35:28 crc kubenswrapper[4904]: I0214 12:35:28.128236 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad6c70dd-1fb9-416c-aebf-15af183a1b4c" containerName="extract-content" Feb 14 12:35:28 crc kubenswrapper[4904]: E0214 12:35:28.128331 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad6c70dd-1fb9-416c-aebf-15af183a1b4c" containerName="extract-utilities" Feb 14 12:35:28 crc kubenswrapper[4904]: I0214 12:35:28.128396 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad6c70dd-1fb9-416c-aebf-15af183a1b4c" containerName="extract-utilities" Feb 14 12:35:28 crc kubenswrapper[4904]: E0214 12:35:28.128463 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad6c70dd-1fb9-416c-aebf-15af183a1b4c" containerName="registry-server" Feb 14 12:35:28 crc kubenswrapper[4904]: I0214 12:35:28.128521 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad6c70dd-1fb9-416c-aebf-15af183a1b4c" containerName="registry-server" Feb 14 12:35:28 crc kubenswrapper[4904]: I0214 12:35:28.128775 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad6c70dd-1fb9-416c-aebf-15af183a1b4c" containerName="registry-server" Feb 14 12:35:28 crc kubenswrapper[4904]: I0214 12:35:28.130120 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-98pb6" Feb 14 12:35:28 crc kubenswrapper[4904]: I0214 12:35:28.148928 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-98pb6"] Feb 14 12:35:28 crc kubenswrapper[4904]: I0214 12:35:28.240750 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7sjd\" (UniqueName: \"kubernetes.io/projected/69668d5f-e6a3-4dac-b8ca-51644c0b7285-kube-api-access-p7sjd\") pod \"certified-operators-98pb6\" (UID: \"69668d5f-e6a3-4dac-b8ca-51644c0b7285\") " pod="openshift-marketplace/certified-operators-98pb6" Feb 14 12:35:28 crc kubenswrapper[4904]: I0214 12:35:28.241060 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69668d5f-e6a3-4dac-b8ca-51644c0b7285-utilities\") pod \"certified-operators-98pb6\" (UID: \"69668d5f-e6a3-4dac-b8ca-51644c0b7285\") " pod="openshift-marketplace/certified-operators-98pb6" Feb 14 12:35:28 crc kubenswrapper[4904]: I0214 12:35:28.241253 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69668d5f-e6a3-4dac-b8ca-51644c0b7285-catalog-content\") pod \"certified-operators-98pb6\" (UID: \"69668d5f-e6a3-4dac-b8ca-51644c0b7285\") " pod="openshift-marketplace/certified-operators-98pb6" Feb 14 12:35:28 crc kubenswrapper[4904]: I0214 12:35:28.343330 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69668d5f-e6a3-4dac-b8ca-51644c0b7285-utilities\") pod \"certified-operators-98pb6\" (UID: \"69668d5f-e6a3-4dac-b8ca-51644c0b7285\") " pod="openshift-marketplace/certified-operators-98pb6" Feb 14 12:35:28 crc kubenswrapper[4904]: I0214 12:35:28.343420 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69668d5f-e6a3-4dac-b8ca-51644c0b7285-catalog-content\") pod \"certified-operators-98pb6\" (UID: \"69668d5f-e6a3-4dac-b8ca-51644c0b7285\") " pod="openshift-marketplace/certified-operators-98pb6" Feb 14 12:35:28 crc kubenswrapper[4904]: I0214 12:35:28.343496 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7sjd\" (UniqueName: \"kubernetes.io/projected/69668d5f-e6a3-4dac-b8ca-51644c0b7285-kube-api-access-p7sjd\") pod \"certified-operators-98pb6\" (UID: \"69668d5f-e6a3-4dac-b8ca-51644c0b7285\") " pod="openshift-marketplace/certified-operators-98pb6" Feb 14 12:35:28 crc kubenswrapper[4904]: I0214 12:35:28.343771 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69668d5f-e6a3-4dac-b8ca-51644c0b7285-utilities\") pod \"certified-operators-98pb6\" (UID: \"69668d5f-e6a3-4dac-b8ca-51644c0b7285\") " pod="openshift-marketplace/certified-operators-98pb6" Feb 14 12:35:28 crc kubenswrapper[4904]: I0214 12:35:28.343874 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69668d5f-e6a3-4dac-b8ca-51644c0b7285-catalog-content\") pod \"certified-operators-98pb6\" (UID: \"69668d5f-e6a3-4dac-b8ca-51644c0b7285\") " pod="openshift-marketplace/certified-operators-98pb6" Feb 14 12:35:28 crc kubenswrapper[4904]: I0214 12:35:28.367741 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7sjd\" (UniqueName: \"kubernetes.io/projected/69668d5f-e6a3-4dac-b8ca-51644c0b7285-kube-api-access-p7sjd\") pod \"certified-operators-98pb6\" (UID: \"69668d5f-e6a3-4dac-b8ca-51644c0b7285\") " pod="openshift-marketplace/certified-operators-98pb6" Feb 14 12:35:28 crc kubenswrapper[4904]: I0214 12:35:28.451263 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-98pb6" Feb 14 12:35:29 crc kubenswrapper[4904]: I0214 12:35:29.020223 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-98pb6"] Feb 14 12:35:29 crc kubenswrapper[4904]: I0214 12:35:29.206402 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98pb6" event={"ID":"69668d5f-e6a3-4dac-b8ca-51644c0b7285","Type":"ContainerStarted","Data":"49f5cf5130163fb52c498ba9d343d792973b9d92d6a24ca80324a2c28d9a3f1e"} Feb 14 12:35:30 crc kubenswrapper[4904]: I0214 12:35:30.223546 4904 generic.go:334] "Generic (PLEG): container finished" podID="69668d5f-e6a3-4dac-b8ca-51644c0b7285" containerID="14c2b616abb874b118bc0caa67b826bb48742cca8ebbe07b45ca97e87fce1a0a" exitCode=0 Feb 14 12:35:30 crc kubenswrapper[4904]: I0214 12:35:30.223722 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98pb6" event={"ID":"69668d5f-e6a3-4dac-b8ca-51644c0b7285","Type":"ContainerDied","Data":"14c2b616abb874b118bc0caa67b826bb48742cca8ebbe07b45ca97e87fce1a0a"} Feb 14 12:35:30 crc kubenswrapper[4904]: I0214 12:35:30.969651 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t_b4eac63f-c406-48a9-ac8b-6150cece2deb/util/0.log" Feb 14 12:35:31 crc kubenswrapper[4904]: I0214 12:35:31.205196 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t_b4eac63f-c406-48a9-ac8b-6150cece2deb/util/0.log" Feb 14 12:35:31 crc kubenswrapper[4904]: I0214 12:35:31.212425 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t_b4eac63f-c406-48a9-ac8b-6150cece2deb/pull/0.log" Feb 14 12:35:31 crc kubenswrapper[4904]: I0214 12:35:31.232305 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98pb6" event={"ID":"69668d5f-e6a3-4dac-b8ca-51644c0b7285","Type":"ContainerStarted","Data":"1b352c0617ba5ef5b8aff0ce15754cf4656fac57a6b3ca2b6806685f24a42a36"} Feb 14 12:35:31 crc kubenswrapper[4904]: I0214 12:35:31.271363 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t_b4eac63f-c406-48a9-ac8b-6150cece2deb/pull/0.log" Feb 14 12:35:31 crc kubenswrapper[4904]: I0214 12:35:31.460386 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t_b4eac63f-c406-48a9-ac8b-6150cece2deb/extract/0.log" Feb 14 12:35:31 crc kubenswrapper[4904]: I0214 12:35:31.469096 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t_b4eac63f-c406-48a9-ac8b-6150cece2deb/util/0.log" Feb 14 12:35:31 crc kubenswrapper[4904]: I0214 12:35:31.488586 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213s277t_b4eac63f-c406-48a9-ac8b-6150cece2deb/pull/0.log" Feb 14 12:35:31 crc kubenswrapper[4904]: I0214 12:35:31.671037 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4m6zf_dfb9b35f-85bb-43ec-a29d-897b3ebf0da9/extract-utilities/0.log" Feb 14 12:35:31 crc kubenswrapper[4904]: I0214 12:35:31.818290 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4m6zf_dfb9b35f-85bb-43ec-a29d-897b3ebf0da9/extract-utilities/0.log" Feb 14 12:35:31 crc kubenswrapper[4904]: I0214 12:35:31.880601 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4m6zf_dfb9b35f-85bb-43ec-a29d-897b3ebf0da9/extract-content/0.log" Feb 14 12:35:31 crc kubenswrapper[4904]: I0214 12:35:31.889184 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4m6zf_dfb9b35f-85bb-43ec-a29d-897b3ebf0da9/extract-content/0.log" Feb 14 12:35:32 crc kubenswrapper[4904]: I0214 12:35:32.073024 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4m6zf_dfb9b35f-85bb-43ec-a29d-897b3ebf0da9/extract-content/0.log" Feb 14 12:35:32 crc kubenswrapper[4904]: I0214 12:35:32.084656 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4m6zf_dfb9b35f-85bb-43ec-a29d-897b3ebf0da9/extract-utilities/0.log" Feb 14 12:35:32 crc kubenswrapper[4904]: I0214 12:35:32.247618 4904 generic.go:334] "Generic (PLEG): container finished" podID="69668d5f-e6a3-4dac-b8ca-51644c0b7285" containerID="1b352c0617ba5ef5b8aff0ce15754cf4656fac57a6b3ca2b6806685f24a42a36" exitCode=0 Feb 14 12:35:32 crc kubenswrapper[4904]: I0214 12:35:32.247654 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98pb6" event={"ID":"69668d5f-e6a3-4dac-b8ca-51644c0b7285","Type":"ContainerDied","Data":"1b352c0617ba5ef5b8aff0ce15754cf4656fac57a6b3ca2b6806685f24a42a36"} Feb 14 12:35:32 crc kubenswrapper[4904]: I0214 12:35:32.290754 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-98pb6_69668d5f-e6a3-4dac-b8ca-51644c0b7285/extract-utilities/0.log" Feb 14 12:35:32 crc kubenswrapper[4904]: I0214 12:35:32.662053 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-98pb6_69668d5f-e6a3-4dac-b8ca-51644c0b7285/extract-content/0.log" Feb 14 12:35:32 crc kubenswrapper[4904]: I0214 12:35:32.744021 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-98pb6_69668d5f-e6a3-4dac-b8ca-51644c0b7285/extract-utilities/0.log" Feb 14 12:35:32 crc kubenswrapper[4904]: I0214 12:35:32.744111 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-98pb6_69668d5f-e6a3-4dac-b8ca-51644c0b7285/extract-content/0.log" Feb 14 12:35:32 crc kubenswrapper[4904]: I0214 12:35:32.783383 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4m6zf_dfb9b35f-85bb-43ec-a29d-897b3ebf0da9/registry-server/0.log" Feb 14 12:35:32 crc kubenswrapper[4904]: I0214 12:35:32.971885 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-98pb6_69668d5f-e6a3-4dac-b8ca-51644c0b7285/extract-content/0.log" Feb 14 12:35:33 crc kubenswrapper[4904]: I0214 12:35:33.003138 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-98pb6_69668d5f-e6a3-4dac-b8ca-51644c0b7285/extract-utilities/0.log" Feb 14 12:35:33 crc kubenswrapper[4904]: I0214 12:35:33.243524 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-99bfv_4233cb32-d700-4c17-a97b-82d5bcd6ea5f/extract-utilities/0.log" Feb 14 12:35:33 crc kubenswrapper[4904]: I0214 12:35:33.257801 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98pb6" event={"ID":"69668d5f-e6a3-4dac-b8ca-51644c0b7285","Type":"ContainerStarted","Data":"fe10206f16e8efbf633bb7f7876114a338872131ab1982338f6b1b5449c5478a"} Feb 14 12:35:33 crc kubenswrapper[4904]: I0214 12:35:33.278049 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-98pb6" podStartSLOduration=2.841198027 podStartE2EDuration="5.278034337s" podCreationTimestamp="2026-02-14 12:35:28 +0000 UTC" firstStartedPulling="2026-02-14 12:35:30.227775317 +0000 UTC m=+5121.040539978" lastFinishedPulling="2026-02-14 12:35:32.664611627 +0000 UTC m=+5123.477376288" observedRunningTime="2026-02-14 12:35:33.272485625 +0000 UTC m=+5124.085250286" watchObservedRunningTime="2026-02-14 12:35:33.278034337 +0000 UTC m=+5124.090798998" Feb 14 12:35:33 crc kubenswrapper[4904]: I0214 12:35:33.489470 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-99bfv_4233cb32-d700-4c17-a97b-82d5bcd6ea5f/extract-utilities/0.log" Feb 14 12:35:33 crc kubenswrapper[4904]: I0214 12:35:33.544216 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-99bfv_4233cb32-d700-4c17-a97b-82d5bcd6ea5f/extract-content/0.log" Feb 14 12:35:33 crc kubenswrapper[4904]: I0214 12:35:33.583857 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-99bfv_4233cb32-d700-4c17-a97b-82d5bcd6ea5f/extract-content/0.log" Feb 14 12:35:33 crc kubenswrapper[4904]: I0214 12:35:33.774752 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-99bfv_4233cb32-d700-4c17-a97b-82d5bcd6ea5f/extract-utilities/0.log" Feb 14 12:35:33 crc kubenswrapper[4904]: I0214 12:35:33.839942 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:35:33 crc kubenswrapper[4904]: E0214 12:35:33.840199 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:35:33 crc kubenswrapper[4904]: I0214 12:35:33.901148 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-99bfv_4233cb32-d700-4c17-a97b-82d5bcd6ea5f/extract-content/0.log" Feb 14 12:35:34 crc kubenswrapper[4904]: I0214 12:35:34.077714 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn_becbe5d6-16f4-4f62-8c90-70be78f0c091/util/0.log" Feb 14 12:35:34 crc kubenswrapper[4904]: I0214 12:35:34.540292 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-99bfv_4233cb32-d700-4c17-a97b-82d5bcd6ea5f/registry-server/0.log" Feb 14 12:35:34 crc kubenswrapper[4904]: I0214 12:35:34.565531 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn_becbe5d6-16f4-4f62-8c90-70be78f0c091/util/0.log" Feb 14 12:35:34 crc kubenswrapper[4904]: I0214 12:35:34.576589 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn_becbe5d6-16f4-4f62-8c90-70be78f0c091/pull/0.log" Feb 14 12:35:34 crc kubenswrapper[4904]: I0214 12:35:34.621927 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn_becbe5d6-16f4-4f62-8c90-70be78f0c091/pull/0.log" Feb 14 12:35:34 crc kubenswrapper[4904]: I0214 12:35:34.753343 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn_becbe5d6-16f4-4f62-8c90-70be78f0c091/util/0.log" Feb 14 12:35:34 crc kubenswrapper[4904]: I0214 12:35:34.783100 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn_becbe5d6-16f4-4f62-8c90-70be78f0c091/pull/0.log" Feb 14 12:35:34 crc kubenswrapper[4904]: I0214 12:35:34.851228 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323eca2t4kn_becbe5d6-16f4-4f62-8c90-70be78f0c091/extract/0.log" Feb 14 12:35:34 crc kubenswrapper[4904]: I0214 12:35:34.929491 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-h5v26_2104a707-8d9b-43e6-9c74-f41bdafcf818/marketplace-operator/0.log" Feb 14 12:35:35 crc kubenswrapper[4904]: I0214 12:35:35.125545 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h2tbz_9a1b8061-2687-4083-8e86-6c9b99daf8a5/extract-utilities/0.log" Feb 14 12:35:35 crc kubenswrapper[4904]: I0214 12:35:35.315617 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h2tbz_9a1b8061-2687-4083-8e86-6c9b99daf8a5/extract-content/0.log" Feb 14 12:35:35 crc kubenswrapper[4904]: I0214 12:35:35.325581 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h2tbz_9a1b8061-2687-4083-8e86-6c9b99daf8a5/extract-content/0.log" Feb 14 12:35:35 crc kubenswrapper[4904]: I0214 12:35:35.335676 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h2tbz_9a1b8061-2687-4083-8e86-6c9b99daf8a5/extract-utilities/0.log" Feb 14 12:35:35 crc kubenswrapper[4904]: I0214 12:35:35.591198 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h2tbz_9a1b8061-2687-4083-8e86-6c9b99daf8a5/extract-content/0.log" Feb 14 12:35:35 crc kubenswrapper[4904]: I0214 12:35:35.622136 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h2tbz_9a1b8061-2687-4083-8e86-6c9b99daf8a5/extract-utilities/0.log" Feb 14 12:35:35 crc kubenswrapper[4904]: I0214 12:35:35.671046 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h2tbz_9a1b8061-2687-4083-8e86-6c9b99daf8a5/registry-server/0.log" Feb 14 12:35:35 crc kubenswrapper[4904]: I0214 12:35:35.771194 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wxxxn_6b349538-4188-447c-aee9-74797e324398/extract-utilities/0.log" Feb 14 12:35:35 crc kubenswrapper[4904]: I0214 12:35:35.936577 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wxxxn_6b349538-4188-447c-aee9-74797e324398/extract-content/0.log" Feb 14 12:35:35 crc kubenswrapper[4904]: I0214 12:35:35.940102 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wxxxn_6b349538-4188-447c-aee9-74797e324398/extract-utilities/0.log" Feb 14 12:35:35 crc kubenswrapper[4904]: I0214 12:35:35.957901 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wxxxn_6b349538-4188-447c-aee9-74797e324398/extract-content/0.log" Feb 14 12:35:36 crc kubenswrapper[4904]: I0214 12:35:36.134336 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wxxxn_6b349538-4188-447c-aee9-74797e324398/extract-content/0.log" Feb 14 12:35:36 crc kubenswrapper[4904]: I0214 12:35:36.160920 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wxxxn_6b349538-4188-447c-aee9-74797e324398/extract-utilities/0.log" Feb 14 12:35:36 crc kubenswrapper[4904]: I0214 12:35:36.828607 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-wxxxn_6b349538-4188-447c-aee9-74797e324398/registry-server/0.log" Feb 14 12:35:38 crc kubenswrapper[4904]: I0214 12:35:38.452538 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-98pb6" Feb 14 12:35:38 crc kubenswrapper[4904]: I0214 12:35:38.453420 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-98pb6" Feb 14 12:35:38 crc kubenswrapper[4904]: I0214 12:35:38.510205 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-98pb6" Feb 14 12:35:39 crc kubenswrapper[4904]: I0214 12:35:39.364940 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-98pb6" Feb 14 12:35:39 crc kubenswrapper[4904]: I0214 12:35:39.434330 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-98pb6"] Feb 14 12:35:41 crc kubenswrapper[4904]: I0214 12:35:41.321327 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-98pb6" podUID="69668d5f-e6a3-4dac-b8ca-51644c0b7285" containerName="registry-server" containerID="cri-o://fe10206f16e8efbf633bb7f7876114a338872131ab1982338f6b1b5449c5478a" gracePeriod=2 Feb 14 12:35:41 crc kubenswrapper[4904]: I0214 12:35:41.848812 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-98pb6" Feb 14 12:35:41 crc kubenswrapper[4904]: I0214 12:35:41.986262 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7sjd\" (UniqueName: \"kubernetes.io/projected/69668d5f-e6a3-4dac-b8ca-51644c0b7285-kube-api-access-p7sjd\") pod \"69668d5f-e6a3-4dac-b8ca-51644c0b7285\" (UID: \"69668d5f-e6a3-4dac-b8ca-51644c0b7285\") " Feb 14 12:35:41 crc kubenswrapper[4904]: I0214 12:35:41.986368 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69668d5f-e6a3-4dac-b8ca-51644c0b7285-catalog-content\") pod \"69668d5f-e6a3-4dac-b8ca-51644c0b7285\" (UID: \"69668d5f-e6a3-4dac-b8ca-51644c0b7285\") " Feb 14 12:35:41 crc kubenswrapper[4904]: I0214 12:35:41.986440 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69668d5f-e6a3-4dac-b8ca-51644c0b7285-utilities\") pod \"69668d5f-e6a3-4dac-b8ca-51644c0b7285\" (UID: \"69668d5f-e6a3-4dac-b8ca-51644c0b7285\") " Feb 14 12:35:41 crc kubenswrapper[4904]: I0214 12:35:41.987206 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69668d5f-e6a3-4dac-b8ca-51644c0b7285-utilities" (OuterVolumeSpecName: "utilities") pod "69668d5f-e6a3-4dac-b8ca-51644c0b7285" (UID: "69668d5f-e6a3-4dac-b8ca-51644c0b7285"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:35:41 crc kubenswrapper[4904]: I0214 12:35:41.989225 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69668d5f-e6a3-4dac-b8ca-51644c0b7285-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 12:35:41 crc kubenswrapper[4904]: I0214 12:35:41.992581 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69668d5f-e6a3-4dac-b8ca-51644c0b7285-kube-api-access-p7sjd" (OuterVolumeSpecName: "kube-api-access-p7sjd") pod "69668d5f-e6a3-4dac-b8ca-51644c0b7285" (UID: "69668d5f-e6a3-4dac-b8ca-51644c0b7285"). InnerVolumeSpecName "kube-api-access-p7sjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:35:42 crc kubenswrapper[4904]: I0214 12:35:42.039797 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69668d5f-e6a3-4dac-b8ca-51644c0b7285-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69668d5f-e6a3-4dac-b8ca-51644c0b7285" (UID: "69668d5f-e6a3-4dac-b8ca-51644c0b7285"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:35:42 crc kubenswrapper[4904]: I0214 12:35:42.091818 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7sjd\" (UniqueName: \"kubernetes.io/projected/69668d5f-e6a3-4dac-b8ca-51644c0b7285-kube-api-access-p7sjd\") on node \"crc\" DevicePath \"\"" Feb 14 12:35:42 crc kubenswrapper[4904]: I0214 12:35:42.091862 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69668d5f-e6a3-4dac-b8ca-51644c0b7285-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 12:35:42 crc kubenswrapper[4904]: I0214 12:35:42.330017 4904 generic.go:334] "Generic (PLEG): container finished" podID="69668d5f-e6a3-4dac-b8ca-51644c0b7285" containerID="fe10206f16e8efbf633bb7f7876114a338872131ab1982338f6b1b5449c5478a" exitCode=0 Feb 14 12:35:42 crc kubenswrapper[4904]: I0214 12:35:42.330059 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98pb6" event={"ID":"69668d5f-e6a3-4dac-b8ca-51644c0b7285","Type":"ContainerDied","Data":"fe10206f16e8efbf633bb7f7876114a338872131ab1982338f6b1b5449c5478a"} Feb 14 12:35:42 crc kubenswrapper[4904]: I0214 12:35:42.330086 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98pb6" event={"ID":"69668d5f-e6a3-4dac-b8ca-51644c0b7285","Type":"ContainerDied","Data":"49f5cf5130163fb52c498ba9d343d792973b9d92d6a24ca80324a2c28d9a3f1e"} Feb 14 12:35:42 crc kubenswrapper[4904]: I0214 12:35:42.330103 4904 scope.go:117] "RemoveContainer" containerID="fe10206f16e8efbf633bb7f7876114a338872131ab1982338f6b1b5449c5478a" Feb 14 12:35:42 crc kubenswrapper[4904]: I0214 12:35:42.330110 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-98pb6" Feb 14 12:35:42 crc kubenswrapper[4904]: I0214 12:35:42.350387 4904 scope.go:117] "RemoveContainer" containerID="1b352c0617ba5ef5b8aff0ce15754cf4656fac57a6b3ca2b6806685f24a42a36" Feb 14 12:35:42 crc kubenswrapper[4904]: I0214 12:35:42.372272 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-98pb6"] Feb 14 12:35:42 crc kubenswrapper[4904]: I0214 12:35:42.380488 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-98pb6"] Feb 14 12:35:42 crc kubenswrapper[4904]: I0214 12:35:42.386451 4904 scope.go:117] "RemoveContainer" containerID="14c2b616abb874b118bc0caa67b826bb48742cca8ebbe07b45ca97e87fce1a0a" Feb 14 12:35:42 crc kubenswrapper[4904]: I0214 12:35:42.438620 4904 scope.go:117] "RemoveContainer" containerID="fe10206f16e8efbf633bb7f7876114a338872131ab1982338f6b1b5449c5478a" Feb 14 12:35:42 crc kubenswrapper[4904]: E0214 12:35:42.439302 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe10206f16e8efbf633bb7f7876114a338872131ab1982338f6b1b5449c5478a\": container with ID starting with fe10206f16e8efbf633bb7f7876114a338872131ab1982338f6b1b5449c5478a not found: ID does not exist" containerID="fe10206f16e8efbf633bb7f7876114a338872131ab1982338f6b1b5449c5478a" Feb 14 12:35:42 crc kubenswrapper[4904]: I0214 12:35:42.439561 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe10206f16e8efbf633bb7f7876114a338872131ab1982338f6b1b5449c5478a"} err="failed to get container status \"fe10206f16e8efbf633bb7f7876114a338872131ab1982338f6b1b5449c5478a\": rpc error: code = NotFound desc = could not find container \"fe10206f16e8efbf633bb7f7876114a338872131ab1982338f6b1b5449c5478a\": container with ID starting with fe10206f16e8efbf633bb7f7876114a338872131ab1982338f6b1b5449c5478a not found: ID does not exist" Feb 14 12:35:42 crc kubenswrapper[4904]: I0214 12:35:42.440013 4904 scope.go:117] "RemoveContainer" containerID="1b352c0617ba5ef5b8aff0ce15754cf4656fac57a6b3ca2b6806685f24a42a36" Feb 14 12:35:42 crc kubenswrapper[4904]: E0214 12:35:42.440847 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b352c0617ba5ef5b8aff0ce15754cf4656fac57a6b3ca2b6806685f24a42a36\": container with ID starting with 1b352c0617ba5ef5b8aff0ce15754cf4656fac57a6b3ca2b6806685f24a42a36 not found: ID does not exist" containerID="1b352c0617ba5ef5b8aff0ce15754cf4656fac57a6b3ca2b6806685f24a42a36" Feb 14 12:35:42 crc kubenswrapper[4904]: I0214 12:35:42.440905 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b352c0617ba5ef5b8aff0ce15754cf4656fac57a6b3ca2b6806685f24a42a36"} err="failed to get container status \"1b352c0617ba5ef5b8aff0ce15754cf4656fac57a6b3ca2b6806685f24a42a36\": rpc error: code = NotFound desc = could not find container \"1b352c0617ba5ef5b8aff0ce15754cf4656fac57a6b3ca2b6806685f24a42a36\": container with ID starting with 1b352c0617ba5ef5b8aff0ce15754cf4656fac57a6b3ca2b6806685f24a42a36 not found: ID does not exist" Feb 14 12:35:42 crc kubenswrapper[4904]: I0214 12:35:42.440931 4904 scope.go:117] "RemoveContainer" containerID="14c2b616abb874b118bc0caa67b826bb48742cca8ebbe07b45ca97e87fce1a0a" Feb 14 12:35:42 crc kubenswrapper[4904]: E0214 12:35:42.441431 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14c2b616abb874b118bc0caa67b826bb48742cca8ebbe07b45ca97e87fce1a0a\": container with ID starting with 14c2b616abb874b118bc0caa67b826bb48742cca8ebbe07b45ca97e87fce1a0a not found: ID does not exist" containerID="14c2b616abb874b118bc0caa67b826bb48742cca8ebbe07b45ca97e87fce1a0a" Feb 14 12:35:42 crc kubenswrapper[4904]: I0214 12:35:42.441454 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14c2b616abb874b118bc0caa67b826bb48742cca8ebbe07b45ca97e87fce1a0a"} err="failed to get container status \"14c2b616abb874b118bc0caa67b826bb48742cca8ebbe07b45ca97e87fce1a0a\": rpc error: code = NotFound desc = could not find container \"14c2b616abb874b118bc0caa67b826bb48742cca8ebbe07b45ca97e87fce1a0a\": container with ID starting with 14c2b616abb874b118bc0caa67b826bb48742cca8ebbe07b45ca97e87fce1a0a not found: ID does not exist" Feb 14 12:35:43 crc kubenswrapper[4904]: I0214 12:35:43.849494 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69668d5f-e6a3-4dac-b8ca-51644c0b7285" path="/var/lib/kubelet/pods/69668d5f-e6a3-4dac-b8ca-51644c0b7285/volumes" Feb 14 12:35:47 crc kubenswrapper[4904]: I0214 12:35:47.839650 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:35:47 crc kubenswrapper[4904]: E0214 12:35:47.841396 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:36:00 crc kubenswrapper[4904]: I0214 12:36:00.836067 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:36:00 crc kubenswrapper[4904]: E0214 12:36:00.837008 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:36:12 crc kubenswrapper[4904]: I0214 12:36:12.835948 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:36:12 crc kubenswrapper[4904]: E0214 12:36:12.836643 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:36:27 crc kubenswrapper[4904]: I0214 12:36:27.838476 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:36:27 crc kubenswrapper[4904]: E0214 12:36:27.840612 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:36:38 crc kubenswrapper[4904]: I0214 12:36:38.837609 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:36:38 crc kubenswrapper[4904]: E0214 12:36:38.839495 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:36:50 crc kubenswrapper[4904]: I0214 12:36:50.836248 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:36:50 crc kubenswrapper[4904]: E0214 12:36:50.836796 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:37:03 crc kubenswrapper[4904]: I0214 12:37:03.836148 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:37:03 crc kubenswrapper[4904]: E0214 12:37:03.837040 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:37:18 crc kubenswrapper[4904]: I0214 12:37:18.836496 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:37:18 crc kubenswrapper[4904]: E0214 12:37:18.837212 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:37:30 crc kubenswrapper[4904]: I0214 12:37:30.837366 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:37:30 crc kubenswrapper[4904]: E0214 12:37:30.838499 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:37:35 crc kubenswrapper[4904]: I0214 12:37:35.339885 4904 generic.go:334] "Generic (PLEG): container finished" podID="2283857f-6998-4d33-b42b-40c360586a64" containerID="e71823cf89cf36c0c1b4707d050d4ac82a47cdc22a3061e00c700a665312eee1" exitCode=0 Feb 14 12:37:35 crc kubenswrapper[4904]: I0214 12:37:35.341237 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-whwk6/must-gather-cdvfw" event={"ID":"2283857f-6998-4d33-b42b-40c360586a64","Type":"ContainerDied","Data":"e71823cf89cf36c0c1b4707d050d4ac82a47cdc22a3061e00c700a665312eee1"} Feb 14 12:37:35 crc kubenswrapper[4904]: I0214 12:37:35.343545 4904 scope.go:117] "RemoveContainer" containerID="e71823cf89cf36c0c1b4707d050d4ac82a47cdc22a3061e00c700a665312eee1" Feb 14 12:37:35 crc kubenswrapper[4904]: I0214 12:37:35.611062 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-whwk6_must-gather-cdvfw_2283857f-6998-4d33-b42b-40c360586a64/gather/0.log" Feb 14 12:37:41 crc kubenswrapper[4904]: I0214 12:37:41.836712 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:37:41 crc kubenswrapper[4904]: E0214 12:37:41.837405 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:37:43 crc kubenswrapper[4904]: I0214 12:37:43.515087 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-whwk6/must-gather-cdvfw"] Feb 14 12:37:43 crc kubenswrapper[4904]: I0214 12:37:43.515823 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-whwk6/must-gather-cdvfw" podUID="2283857f-6998-4d33-b42b-40c360586a64" containerName="copy" containerID="cri-o://75dd94c86c2a31e73d5e0986164419e967c3c97b54727d6290e89687501680ad" gracePeriod=2 Feb 14 12:37:43 crc kubenswrapper[4904]: I0214 12:37:43.524576 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-whwk6/must-gather-cdvfw"] Feb 14 12:37:43 crc kubenswrapper[4904]: I0214 12:37:43.937194 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-whwk6_must-gather-cdvfw_2283857f-6998-4d33-b42b-40c360586a64/copy/0.log" Feb 14 12:37:43 crc kubenswrapper[4904]: I0214 12:37:43.938644 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-whwk6/must-gather-cdvfw" Feb 14 12:37:44 crc kubenswrapper[4904]: I0214 12:37:44.071397 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2283857f-6998-4d33-b42b-40c360586a64-must-gather-output\") pod \"2283857f-6998-4d33-b42b-40c360586a64\" (UID: \"2283857f-6998-4d33-b42b-40c360586a64\") " Feb 14 12:37:44 crc kubenswrapper[4904]: I0214 12:37:44.071498 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99bbv\" (UniqueName: \"kubernetes.io/projected/2283857f-6998-4d33-b42b-40c360586a64-kube-api-access-99bbv\") pod \"2283857f-6998-4d33-b42b-40c360586a64\" (UID: \"2283857f-6998-4d33-b42b-40c360586a64\") " Feb 14 12:37:44 crc kubenswrapper[4904]: I0214 12:37:44.078151 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2283857f-6998-4d33-b42b-40c360586a64-kube-api-access-99bbv" (OuterVolumeSpecName: "kube-api-access-99bbv") pod "2283857f-6998-4d33-b42b-40c360586a64" (UID: "2283857f-6998-4d33-b42b-40c360586a64"). InnerVolumeSpecName "kube-api-access-99bbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:37:44 crc kubenswrapper[4904]: I0214 12:37:44.173338 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99bbv\" (UniqueName: \"kubernetes.io/projected/2283857f-6998-4d33-b42b-40c360586a64-kube-api-access-99bbv\") on node \"crc\" DevicePath \"\"" Feb 14 12:37:44 crc kubenswrapper[4904]: I0214 12:37:44.221693 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2283857f-6998-4d33-b42b-40c360586a64-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "2283857f-6998-4d33-b42b-40c360586a64" (UID: "2283857f-6998-4d33-b42b-40c360586a64"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:37:44 crc kubenswrapper[4904]: I0214 12:37:44.274993 4904 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2283857f-6998-4d33-b42b-40c360586a64-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 14 12:37:44 crc kubenswrapper[4904]: I0214 12:37:44.416253 4904 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-whwk6_must-gather-cdvfw_2283857f-6998-4d33-b42b-40c360586a64/copy/0.log" Feb 14 12:37:44 crc kubenswrapper[4904]: I0214 12:37:44.417029 4904 generic.go:334] "Generic (PLEG): container finished" podID="2283857f-6998-4d33-b42b-40c360586a64" containerID="75dd94c86c2a31e73d5e0986164419e967c3c97b54727d6290e89687501680ad" exitCode=143 Feb 14 12:37:44 crc kubenswrapper[4904]: I0214 12:37:44.417094 4904 scope.go:117] "RemoveContainer" containerID="75dd94c86c2a31e73d5e0986164419e967c3c97b54727d6290e89687501680ad" Feb 14 12:37:44 crc kubenswrapper[4904]: I0214 12:37:44.417100 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-whwk6/must-gather-cdvfw" Feb 14 12:37:44 crc kubenswrapper[4904]: I0214 12:37:44.440609 4904 scope.go:117] "RemoveContainer" containerID="e71823cf89cf36c0c1b4707d050d4ac82a47cdc22a3061e00c700a665312eee1" Feb 14 12:37:44 crc kubenswrapper[4904]: I0214 12:37:44.584384 4904 scope.go:117] "RemoveContainer" containerID="75dd94c86c2a31e73d5e0986164419e967c3c97b54727d6290e89687501680ad" Feb 14 12:37:44 crc kubenswrapper[4904]: E0214 12:37:44.584870 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75dd94c86c2a31e73d5e0986164419e967c3c97b54727d6290e89687501680ad\": container with ID starting with 75dd94c86c2a31e73d5e0986164419e967c3c97b54727d6290e89687501680ad not found: ID does not exist" containerID="75dd94c86c2a31e73d5e0986164419e967c3c97b54727d6290e89687501680ad" Feb 14 12:37:44 crc kubenswrapper[4904]: I0214 12:37:44.584900 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75dd94c86c2a31e73d5e0986164419e967c3c97b54727d6290e89687501680ad"} err="failed to get container status \"75dd94c86c2a31e73d5e0986164419e967c3c97b54727d6290e89687501680ad\": rpc error: code = NotFound desc = could not find container \"75dd94c86c2a31e73d5e0986164419e967c3c97b54727d6290e89687501680ad\": container with ID starting with 75dd94c86c2a31e73d5e0986164419e967c3c97b54727d6290e89687501680ad not found: ID does not exist" Feb 14 12:37:44 crc kubenswrapper[4904]: I0214 12:37:44.584927 4904 scope.go:117] "RemoveContainer" containerID="e71823cf89cf36c0c1b4707d050d4ac82a47cdc22a3061e00c700a665312eee1" Feb 14 12:37:44 crc kubenswrapper[4904]: E0214 12:37:44.585584 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e71823cf89cf36c0c1b4707d050d4ac82a47cdc22a3061e00c700a665312eee1\": container with ID starting with e71823cf89cf36c0c1b4707d050d4ac82a47cdc22a3061e00c700a665312eee1 not found: ID does not exist" containerID="e71823cf89cf36c0c1b4707d050d4ac82a47cdc22a3061e00c700a665312eee1" Feb 14 12:37:44 crc kubenswrapper[4904]: I0214 12:37:44.585610 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e71823cf89cf36c0c1b4707d050d4ac82a47cdc22a3061e00c700a665312eee1"} err="failed to get container status \"e71823cf89cf36c0c1b4707d050d4ac82a47cdc22a3061e00c700a665312eee1\": rpc error: code = NotFound desc = could not find container \"e71823cf89cf36c0c1b4707d050d4ac82a47cdc22a3061e00c700a665312eee1\": container with ID starting with e71823cf89cf36c0c1b4707d050d4ac82a47cdc22a3061e00c700a665312eee1 not found: ID does not exist" Feb 14 12:37:45 crc kubenswrapper[4904]: I0214 12:37:45.847754 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2283857f-6998-4d33-b42b-40c360586a64" path="/var/lib/kubelet/pods/2283857f-6998-4d33-b42b-40c360586a64/volumes" Feb 14 12:37:55 crc kubenswrapper[4904]: I0214 12:37:55.837477 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:37:55 crc kubenswrapper[4904]: E0214 12:37:55.839079 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:38:10 crc kubenswrapper[4904]: I0214 12:38:10.836685 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:38:10 crc kubenswrapper[4904]: E0214 12:38:10.837351 4904 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-tnf2t_openshift-machine-config-operator(7790d276-2506-4641-8c22-67e6441417f0)\"" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" Feb 14 12:38:14 crc kubenswrapper[4904]: E0214 12:38:14.139963 4904 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/systemd-hostnamed.service\": RecentStats: unable to find data in memory cache]" Feb 14 12:38:23 crc kubenswrapper[4904]: I0214 12:38:23.844227 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" Feb 14 12:38:24 crc kubenswrapper[4904]: I0214 12:38:24.394466 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"46ebab2bf7a2f57a967bd380510e9979599c11a1b684125f1f852ebbd4f4e829"} Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.137454 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nb42q"] Feb 14 12:39:44 crc kubenswrapper[4904]: E0214 12:39:44.138354 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69668d5f-e6a3-4dac-b8ca-51644c0b7285" containerName="extract-utilities" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.138367 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="69668d5f-e6a3-4dac-b8ca-51644c0b7285" containerName="extract-utilities" Feb 14 12:39:44 crc kubenswrapper[4904]: E0214 12:39:44.138387 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69668d5f-e6a3-4dac-b8ca-51644c0b7285" containerName="extract-content" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.138393 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="69668d5f-e6a3-4dac-b8ca-51644c0b7285" containerName="extract-content" Feb 14 12:39:44 crc kubenswrapper[4904]: E0214 12:39:44.138406 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2283857f-6998-4d33-b42b-40c360586a64" containerName="copy" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.138413 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="2283857f-6998-4d33-b42b-40c360586a64" containerName="copy" Feb 14 12:39:44 crc kubenswrapper[4904]: E0214 12:39:44.138431 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69668d5f-e6a3-4dac-b8ca-51644c0b7285" containerName="registry-server" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.138436 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="69668d5f-e6a3-4dac-b8ca-51644c0b7285" containerName="registry-server" Feb 14 12:39:44 crc kubenswrapper[4904]: E0214 12:39:44.138446 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2283857f-6998-4d33-b42b-40c360586a64" containerName="gather" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.138452 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="2283857f-6998-4d33-b42b-40c360586a64" containerName="gather" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.138613 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="2283857f-6998-4d33-b42b-40c360586a64" containerName="gather" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.138625 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="69668d5f-e6a3-4dac-b8ca-51644c0b7285" containerName="registry-server" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.138644 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="2283857f-6998-4d33-b42b-40c360586a64" containerName="copy" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.139894 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nb42q" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.165725 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nb42q"] Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.177052 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fce6644e-da3a-4e27-8c3c-cb79b22598ea-utilities\") pod \"redhat-operators-nb42q\" (UID: \"fce6644e-da3a-4e27-8c3c-cb79b22598ea\") " pod="openshift-marketplace/redhat-operators-nb42q" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.177137 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2khwn\" (UniqueName: \"kubernetes.io/projected/fce6644e-da3a-4e27-8c3c-cb79b22598ea-kube-api-access-2khwn\") pod \"redhat-operators-nb42q\" (UID: \"fce6644e-da3a-4e27-8c3c-cb79b22598ea\") " pod="openshift-marketplace/redhat-operators-nb42q" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.177218 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fce6644e-da3a-4e27-8c3c-cb79b22598ea-catalog-content\") pod \"redhat-operators-nb42q\" (UID: \"fce6644e-da3a-4e27-8c3c-cb79b22598ea\") " pod="openshift-marketplace/redhat-operators-nb42q" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.278391 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fce6644e-da3a-4e27-8c3c-cb79b22598ea-utilities\") pod \"redhat-operators-nb42q\" (UID: \"fce6644e-da3a-4e27-8c3c-cb79b22598ea\") " pod="openshift-marketplace/redhat-operators-nb42q" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.278506 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2khwn\" (UniqueName: \"kubernetes.io/projected/fce6644e-da3a-4e27-8c3c-cb79b22598ea-kube-api-access-2khwn\") pod \"redhat-operators-nb42q\" (UID: \"fce6644e-da3a-4e27-8c3c-cb79b22598ea\") " pod="openshift-marketplace/redhat-operators-nb42q" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.278589 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fce6644e-da3a-4e27-8c3c-cb79b22598ea-catalog-content\") pod \"redhat-operators-nb42q\" (UID: \"fce6644e-da3a-4e27-8c3c-cb79b22598ea\") " pod="openshift-marketplace/redhat-operators-nb42q" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.278857 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fce6644e-da3a-4e27-8c3c-cb79b22598ea-utilities\") pod \"redhat-operators-nb42q\" (UID: \"fce6644e-da3a-4e27-8c3c-cb79b22598ea\") " pod="openshift-marketplace/redhat-operators-nb42q" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.279441 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fce6644e-da3a-4e27-8c3c-cb79b22598ea-catalog-content\") pod \"redhat-operators-nb42q\" (UID: \"fce6644e-da3a-4e27-8c3c-cb79b22598ea\") " pod="openshift-marketplace/redhat-operators-nb42q" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.303964 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2khwn\" (UniqueName: \"kubernetes.io/projected/fce6644e-da3a-4e27-8c3c-cb79b22598ea-kube-api-access-2khwn\") pod \"redhat-operators-nb42q\" (UID: \"fce6644e-da3a-4e27-8c3c-cb79b22598ea\") " pod="openshift-marketplace/redhat-operators-nb42q" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.457760 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nb42q" Feb 14 12:39:44 crc kubenswrapper[4904]: I0214 12:39:44.958328 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nb42q"] Feb 14 12:39:45 crc kubenswrapper[4904]: I0214 12:39:45.126329 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb42q" event={"ID":"fce6644e-da3a-4e27-8c3c-cb79b22598ea","Type":"ContainerStarted","Data":"2e43d820432eb77f42e9f16d5ba740c6475e2cf9c1071cd1a24c665ea178f18e"} Feb 14 12:39:46 crc kubenswrapper[4904]: I0214 12:39:46.134521 4904 generic.go:334] "Generic (PLEG): container finished" podID="fce6644e-da3a-4e27-8c3c-cb79b22598ea" containerID="dc6293173de7058f7a3727730b6caae3e38a6b1beffac6dba16b99ce0021a882" exitCode=0 Feb 14 12:39:46 crc kubenswrapper[4904]: I0214 12:39:46.134573 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb42q" event={"ID":"fce6644e-da3a-4e27-8c3c-cb79b22598ea","Type":"ContainerDied","Data":"dc6293173de7058f7a3727730b6caae3e38a6b1beffac6dba16b99ce0021a882"} Feb 14 12:39:46 crc kubenswrapper[4904]: I0214 12:39:46.136561 4904 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 14 12:39:47 crc kubenswrapper[4904]: I0214 12:39:47.144409 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb42q" event={"ID":"fce6644e-da3a-4e27-8c3c-cb79b22598ea","Type":"ContainerStarted","Data":"c41cbd983d30b88bca26613817071fef4908e209863b74905a986745ca432305"} Feb 14 12:39:52 crc kubenswrapper[4904]: I0214 12:39:52.187234 4904 generic.go:334] "Generic (PLEG): container finished" podID="fce6644e-da3a-4e27-8c3c-cb79b22598ea" containerID="c41cbd983d30b88bca26613817071fef4908e209863b74905a986745ca432305" exitCode=0 Feb 14 12:39:52 crc kubenswrapper[4904]: I0214 12:39:52.187305 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb42q" event={"ID":"fce6644e-da3a-4e27-8c3c-cb79b22598ea","Type":"ContainerDied","Data":"c41cbd983d30b88bca26613817071fef4908e209863b74905a986745ca432305"} Feb 14 12:39:53 crc kubenswrapper[4904]: I0214 12:39:53.202737 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb42q" event={"ID":"fce6644e-da3a-4e27-8c3c-cb79b22598ea","Type":"ContainerStarted","Data":"bde15d6f8385001098dabd62c1d5dc000eb05471b9090b8cea22a8e37120ba52"} Feb 14 12:39:53 crc kubenswrapper[4904]: I0214 12:39:53.221739 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nb42q" podStartSLOduration=2.7084586379999998 podStartE2EDuration="9.221724676s" podCreationTimestamp="2026-02-14 12:39:44 +0000 UTC" firstStartedPulling="2026-02-14 12:39:46.136318873 +0000 UTC m=+5376.949083534" lastFinishedPulling="2026-02-14 12:39:52.649584911 +0000 UTC m=+5383.462349572" observedRunningTime="2026-02-14 12:39:53.220155963 +0000 UTC m=+5384.032920644" watchObservedRunningTime="2026-02-14 12:39:53.221724676 +0000 UTC m=+5384.034489337" Feb 14 12:39:54 crc kubenswrapper[4904]: I0214 12:39:54.459705 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nb42q" Feb 14 12:39:54 crc kubenswrapper[4904]: I0214 12:39:54.460000 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nb42q" Feb 14 12:39:55 crc kubenswrapper[4904]: I0214 12:39:55.513384 4904 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nb42q" podUID="fce6644e-da3a-4e27-8c3c-cb79b22598ea" containerName="registry-server" probeResult="failure" output=< Feb 14 12:39:55 crc kubenswrapper[4904]: timeout: failed to connect service ":50051" within 1s Feb 14 12:39:55 crc kubenswrapper[4904]: > Feb 14 12:40:04 crc kubenswrapper[4904]: I0214 12:40:04.518361 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nb42q" Feb 14 12:40:04 crc kubenswrapper[4904]: I0214 12:40:04.583597 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nb42q" Feb 14 12:40:04 crc kubenswrapper[4904]: I0214 12:40:04.759745 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nb42q"] Feb 14 12:40:06 crc kubenswrapper[4904]: I0214 12:40:06.323210 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nb42q" podUID="fce6644e-da3a-4e27-8c3c-cb79b22598ea" containerName="registry-server" containerID="cri-o://bde15d6f8385001098dabd62c1d5dc000eb05471b9090b8cea22a8e37120ba52" gracePeriod=2 Feb 14 12:40:06 crc kubenswrapper[4904]: I0214 12:40:06.737256 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nb42q" Feb 14 12:40:06 crc kubenswrapper[4904]: I0214 12:40:06.936879 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fce6644e-da3a-4e27-8c3c-cb79b22598ea-utilities\") pod \"fce6644e-da3a-4e27-8c3c-cb79b22598ea\" (UID: \"fce6644e-da3a-4e27-8c3c-cb79b22598ea\") " Feb 14 12:40:06 crc kubenswrapper[4904]: I0214 12:40:06.937207 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fce6644e-da3a-4e27-8c3c-cb79b22598ea-catalog-content\") pod \"fce6644e-da3a-4e27-8c3c-cb79b22598ea\" (UID: \"fce6644e-da3a-4e27-8c3c-cb79b22598ea\") " Feb 14 12:40:06 crc kubenswrapper[4904]: I0214 12:40:06.937341 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2khwn\" (UniqueName: \"kubernetes.io/projected/fce6644e-da3a-4e27-8c3c-cb79b22598ea-kube-api-access-2khwn\") pod \"fce6644e-da3a-4e27-8c3c-cb79b22598ea\" (UID: \"fce6644e-da3a-4e27-8c3c-cb79b22598ea\") " Feb 14 12:40:06 crc kubenswrapper[4904]: I0214 12:40:06.938031 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fce6644e-da3a-4e27-8c3c-cb79b22598ea-utilities" (OuterVolumeSpecName: "utilities") pod "fce6644e-da3a-4e27-8c3c-cb79b22598ea" (UID: "fce6644e-da3a-4e27-8c3c-cb79b22598ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:40:06 crc kubenswrapper[4904]: I0214 12:40:06.945517 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fce6644e-da3a-4e27-8c3c-cb79b22598ea-kube-api-access-2khwn" (OuterVolumeSpecName: "kube-api-access-2khwn") pod "fce6644e-da3a-4e27-8c3c-cb79b22598ea" (UID: "fce6644e-da3a-4e27-8c3c-cb79b22598ea"). InnerVolumeSpecName "kube-api-access-2khwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.039485 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fce6644e-da3a-4e27-8c3c-cb79b22598ea-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.039529 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2khwn\" (UniqueName: \"kubernetes.io/projected/fce6644e-da3a-4e27-8c3c-cb79b22598ea-kube-api-access-2khwn\") on node \"crc\" DevicePath \"\"" Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.061306 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fce6644e-da3a-4e27-8c3c-cb79b22598ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fce6644e-da3a-4e27-8c3c-cb79b22598ea" (UID: "fce6644e-da3a-4e27-8c3c-cb79b22598ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.141613 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fce6644e-da3a-4e27-8c3c-cb79b22598ea-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.331026 4904 generic.go:334] "Generic (PLEG): container finished" podID="fce6644e-da3a-4e27-8c3c-cb79b22598ea" containerID="bde15d6f8385001098dabd62c1d5dc000eb05471b9090b8cea22a8e37120ba52" exitCode=0 Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.331073 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb42q" event={"ID":"fce6644e-da3a-4e27-8c3c-cb79b22598ea","Type":"ContainerDied","Data":"bde15d6f8385001098dabd62c1d5dc000eb05471b9090b8cea22a8e37120ba52"} Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.331122 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb42q" event={"ID":"fce6644e-da3a-4e27-8c3c-cb79b22598ea","Type":"ContainerDied","Data":"2e43d820432eb77f42e9f16d5ba740c6475e2cf9c1071cd1a24c665ea178f18e"} Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.331145 4904 scope.go:117] "RemoveContainer" containerID="bde15d6f8385001098dabd62c1d5dc000eb05471b9090b8cea22a8e37120ba52" Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.331092 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nb42q" Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.361613 4904 scope.go:117] "RemoveContainer" containerID="c41cbd983d30b88bca26613817071fef4908e209863b74905a986745ca432305" Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.369977 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nb42q"] Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.381471 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nb42q"] Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.391427 4904 scope.go:117] "RemoveContainer" containerID="dc6293173de7058f7a3727730b6caae3e38a6b1beffac6dba16b99ce0021a882" Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.422603 4904 scope.go:117] "RemoveContainer" containerID="bde15d6f8385001098dabd62c1d5dc000eb05471b9090b8cea22a8e37120ba52" Feb 14 12:40:07 crc kubenswrapper[4904]: E0214 12:40:07.423240 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bde15d6f8385001098dabd62c1d5dc000eb05471b9090b8cea22a8e37120ba52\": container with ID starting with bde15d6f8385001098dabd62c1d5dc000eb05471b9090b8cea22a8e37120ba52 not found: ID does not exist" containerID="bde15d6f8385001098dabd62c1d5dc000eb05471b9090b8cea22a8e37120ba52" Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.423273 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bde15d6f8385001098dabd62c1d5dc000eb05471b9090b8cea22a8e37120ba52"} err="failed to get container status \"bde15d6f8385001098dabd62c1d5dc000eb05471b9090b8cea22a8e37120ba52\": rpc error: code = NotFound desc = could not find container \"bde15d6f8385001098dabd62c1d5dc000eb05471b9090b8cea22a8e37120ba52\": container with ID starting with bde15d6f8385001098dabd62c1d5dc000eb05471b9090b8cea22a8e37120ba52 not found: ID does not exist" Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.423295 4904 scope.go:117] "RemoveContainer" containerID="c41cbd983d30b88bca26613817071fef4908e209863b74905a986745ca432305" Feb 14 12:40:07 crc kubenswrapper[4904]: E0214 12:40:07.423636 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c41cbd983d30b88bca26613817071fef4908e209863b74905a986745ca432305\": container with ID starting with c41cbd983d30b88bca26613817071fef4908e209863b74905a986745ca432305 not found: ID does not exist" containerID="c41cbd983d30b88bca26613817071fef4908e209863b74905a986745ca432305" Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.423677 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c41cbd983d30b88bca26613817071fef4908e209863b74905a986745ca432305"} err="failed to get container status \"c41cbd983d30b88bca26613817071fef4908e209863b74905a986745ca432305\": rpc error: code = NotFound desc = could not find container \"c41cbd983d30b88bca26613817071fef4908e209863b74905a986745ca432305\": container with ID starting with c41cbd983d30b88bca26613817071fef4908e209863b74905a986745ca432305 not found: ID does not exist" Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.423706 4904 scope.go:117] "RemoveContainer" containerID="dc6293173de7058f7a3727730b6caae3e38a6b1beffac6dba16b99ce0021a882" Feb 14 12:40:07 crc kubenswrapper[4904]: E0214 12:40:07.424036 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc6293173de7058f7a3727730b6caae3e38a6b1beffac6dba16b99ce0021a882\": container with ID starting with dc6293173de7058f7a3727730b6caae3e38a6b1beffac6dba16b99ce0021a882 not found: ID does not exist" containerID="dc6293173de7058f7a3727730b6caae3e38a6b1beffac6dba16b99ce0021a882" Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.424059 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc6293173de7058f7a3727730b6caae3e38a6b1beffac6dba16b99ce0021a882"} err="failed to get container status \"dc6293173de7058f7a3727730b6caae3e38a6b1beffac6dba16b99ce0021a882\": rpc error: code = NotFound desc = could not find container \"dc6293173de7058f7a3727730b6caae3e38a6b1beffac6dba16b99ce0021a882\": container with ID starting with dc6293173de7058f7a3727730b6caae3e38a6b1beffac6dba16b99ce0021a882 not found: ID does not exist" Feb 14 12:40:07 crc kubenswrapper[4904]: I0214 12:40:07.851361 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fce6644e-da3a-4e27-8c3c-cb79b22598ea" path="/var/lib/kubelet/pods/fce6644e-da3a-4e27-8c3c-cb79b22598ea/volumes" Feb 14 12:40:44 crc kubenswrapper[4904]: I0214 12:40:44.229001 4904 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q9fgk"] Feb 14 12:40:44 crc kubenswrapper[4904]: E0214 12:40:44.229802 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fce6644e-da3a-4e27-8c3c-cb79b22598ea" containerName="extract-utilities" Feb 14 12:40:44 crc kubenswrapper[4904]: I0214 12:40:44.229814 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="fce6644e-da3a-4e27-8c3c-cb79b22598ea" containerName="extract-utilities" Feb 14 12:40:44 crc kubenswrapper[4904]: E0214 12:40:44.229880 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fce6644e-da3a-4e27-8c3c-cb79b22598ea" containerName="extract-content" Feb 14 12:40:44 crc kubenswrapper[4904]: I0214 12:40:44.229889 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="fce6644e-da3a-4e27-8c3c-cb79b22598ea" containerName="extract-content" Feb 14 12:40:44 crc kubenswrapper[4904]: E0214 12:40:44.229905 4904 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fce6644e-da3a-4e27-8c3c-cb79b22598ea" containerName="registry-server" Feb 14 12:40:44 crc kubenswrapper[4904]: I0214 12:40:44.229915 4904 state_mem.go:107] "Deleted CPUSet assignment" podUID="fce6644e-da3a-4e27-8c3c-cb79b22598ea" containerName="registry-server" Feb 14 12:40:44 crc kubenswrapper[4904]: I0214 12:40:44.230147 4904 memory_manager.go:354] "RemoveStaleState removing state" podUID="fce6644e-da3a-4e27-8c3c-cb79b22598ea" containerName="registry-server" Feb 14 12:40:44 crc kubenswrapper[4904]: I0214 12:40:44.231373 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q9fgk" Feb 14 12:40:44 crc kubenswrapper[4904]: I0214 12:40:44.247368 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q9fgk"] Feb 14 12:40:44 crc kubenswrapper[4904]: I0214 12:40:44.343770 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn8bz\" (UniqueName: \"kubernetes.io/projected/47680b40-812e-4c78-b77f-52ff4769af25-kube-api-access-nn8bz\") pod \"redhat-marketplace-q9fgk\" (UID: \"47680b40-812e-4c78-b77f-52ff4769af25\") " pod="openshift-marketplace/redhat-marketplace-q9fgk" Feb 14 12:40:44 crc kubenswrapper[4904]: I0214 12:40:44.343983 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47680b40-812e-4c78-b77f-52ff4769af25-utilities\") pod \"redhat-marketplace-q9fgk\" (UID: \"47680b40-812e-4c78-b77f-52ff4769af25\") " pod="openshift-marketplace/redhat-marketplace-q9fgk" Feb 14 12:40:44 crc kubenswrapper[4904]: I0214 12:40:44.344388 4904 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47680b40-812e-4c78-b77f-52ff4769af25-catalog-content\") pod \"redhat-marketplace-q9fgk\" (UID: \"47680b40-812e-4c78-b77f-52ff4769af25\") " pod="openshift-marketplace/redhat-marketplace-q9fgk" Feb 14 12:40:44 crc kubenswrapper[4904]: I0214 12:40:44.445699 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47680b40-812e-4c78-b77f-52ff4769af25-catalog-content\") pod \"redhat-marketplace-q9fgk\" (UID: \"47680b40-812e-4c78-b77f-52ff4769af25\") " pod="openshift-marketplace/redhat-marketplace-q9fgk" Feb 14 12:40:44 crc kubenswrapper[4904]: I0214 12:40:44.445767 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn8bz\" (UniqueName: \"kubernetes.io/projected/47680b40-812e-4c78-b77f-52ff4769af25-kube-api-access-nn8bz\") pod \"redhat-marketplace-q9fgk\" (UID: \"47680b40-812e-4c78-b77f-52ff4769af25\") " pod="openshift-marketplace/redhat-marketplace-q9fgk" Feb 14 12:40:44 crc kubenswrapper[4904]: I0214 12:40:44.445817 4904 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47680b40-812e-4c78-b77f-52ff4769af25-utilities\") pod \"redhat-marketplace-q9fgk\" (UID: \"47680b40-812e-4c78-b77f-52ff4769af25\") " pod="openshift-marketplace/redhat-marketplace-q9fgk" Feb 14 12:40:44 crc kubenswrapper[4904]: I0214 12:40:44.446303 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47680b40-812e-4c78-b77f-52ff4769af25-utilities\") pod \"redhat-marketplace-q9fgk\" (UID: \"47680b40-812e-4c78-b77f-52ff4769af25\") " pod="openshift-marketplace/redhat-marketplace-q9fgk" Feb 14 12:40:44 crc kubenswrapper[4904]: I0214 12:40:44.446309 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47680b40-812e-4c78-b77f-52ff4769af25-catalog-content\") pod \"redhat-marketplace-q9fgk\" (UID: \"47680b40-812e-4c78-b77f-52ff4769af25\") " pod="openshift-marketplace/redhat-marketplace-q9fgk" Feb 14 12:40:44 crc kubenswrapper[4904]: I0214 12:40:44.465760 4904 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn8bz\" (UniqueName: \"kubernetes.io/projected/47680b40-812e-4c78-b77f-52ff4769af25-kube-api-access-nn8bz\") pod \"redhat-marketplace-q9fgk\" (UID: \"47680b40-812e-4c78-b77f-52ff4769af25\") " pod="openshift-marketplace/redhat-marketplace-q9fgk" Feb 14 12:40:44 crc kubenswrapper[4904]: I0214 12:40:44.549819 4904 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q9fgk" Feb 14 12:40:44 crc kubenswrapper[4904]: I0214 12:40:44.870370 4904 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q9fgk"] Feb 14 12:40:45 crc kubenswrapper[4904]: I0214 12:40:45.650711 4904 generic.go:334] "Generic (PLEG): container finished" podID="47680b40-812e-4c78-b77f-52ff4769af25" containerID="1c9b7efc60cae2b059ced6010d5941885723257ff435db2c091617c09b336d78" exitCode=0 Feb 14 12:40:45 crc kubenswrapper[4904]: I0214 12:40:45.650764 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9fgk" event={"ID":"47680b40-812e-4c78-b77f-52ff4769af25","Type":"ContainerDied","Data":"1c9b7efc60cae2b059ced6010d5941885723257ff435db2c091617c09b336d78"} Feb 14 12:40:45 crc kubenswrapper[4904]: I0214 12:40:45.651059 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9fgk" event={"ID":"47680b40-812e-4c78-b77f-52ff4769af25","Type":"ContainerStarted","Data":"46ba01848c723dda8cf0f13c3d56c4dfdc83bcd23622101ec8dfa3b35807fff1"} Feb 14 12:40:46 crc kubenswrapper[4904]: I0214 12:40:46.383418 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:40:46 crc kubenswrapper[4904]: I0214 12:40:46.383704 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:40:46 crc kubenswrapper[4904]: I0214 12:40:46.664751 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9fgk" event={"ID":"47680b40-812e-4c78-b77f-52ff4769af25","Type":"ContainerStarted","Data":"8b101076b67277bf6ea094c1e09e4680362a1dc4ee0ea85ff0dacb49618d5515"} Feb 14 12:40:47 crc kubenswrapper[4904]: I0214 12:40:47.675106 4904 generic.go:334] "Generic (PLEG): container finished" podID="47680b40-812e-4c78-b77f-52ff4769af25" containerID="8b101076b67277bf6ea094c1e09e4680362a1dc4ee0ea85ff0dacb49618d5515" exitCode=0 Feb 14 12:40:47 crc kubenswrapper[4904]: I0214 12:40:47.675459 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9fgk" event={"ID":"47680b40-812e-4c78-b77f-52ff4769af25","Type":"ContainerDied","Data":"8b101076b67277bf6ea094c1e09e4680362a1dc4ee0ea85ff0dacb49618d5515"} Feb 14 12:40:48 crc kubenswrapper[4904]: I0214 12:40:48.684347 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9fgk" event={"ID":"47680b40-812e-4c78-b77f-52ff4769af25","Type":"ContainerStarted","Data":"2250c184f1bb68d4c3f8294a8ea654d620e7ae336e7e0df56c7bdf9baf56356f"} Feb 14 12:40:48 crc kubenswrapper[4904]: I0214 12:40:48.705616 4904 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q9fgk" podStartSLOduration=2.294836952 podStartE2EDuration="4.70559354s" podCreationTimestamp="2026-02-14 12:40:44 +0000 UTC" firstStartedPulling="2026-02-14 12:40:45.652223787 +0000 UTC m=+5436.464988448" lastFinishedPulling="2026-02-14 12:40:48.062980375 +0000 UTC m=+5438.875745036" observedRunningTime="2026-02-14 12:40:48.701493529 +0000 UTC m=+5439.514258190" watchObservedRunningTime="2026-02-14 12:40:48.70559354 +0000 UTC m=+5439.518358201" Feb 14 12:40:54 crc kubenswrapper[4904]: I0214 12:40:54.550021 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q9fgk" Feb 14 12:40:54 crc kubenswrapper[4904]: I0214 12:40:54.550484 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q9fgk" Feb 14 12:40:54 crc kubenswrapper[4904]: I0214 12:40:54.598290 4904 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q9fgk" Feb 14 12:40:54 crc kubenswrapper[4904]: I0214 12:40:54.779644 4904 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q9fgk" Feb 14 12:40:54 crc kubenswrapper[4904]: I0214 12:40:54.848084 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q9fgk"] Feb 14 12:40:56 crc kubenswrapper[4904]: I0214 12:40:56.748030 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q9fgk" podUID="47680b40-812e-4c78-b77f-52ff4769af25" containerName="registry-server" containerID="cri-o://2250c184f1bb68d4c3f8294a8ea654d620e7ae336e7e0df56c7bdf9baf56356f" gracePeriod=2 Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.220211 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q9fgk" Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.281202 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nn8bz\" (UniqueName: \"kubernetes.io/projected/47680b40-812e-4c78-b77f-52ff4769af25-kube-api-access-nn8bz\") pod \"47680b40-812e-4c78-b77f-52ff4769af25\" (UID: \"47680b40-812e-4c78-b77f-52ff4769af25\") " Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.281352 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47680b40-812e-4c78-b77f-52ff4769af25-catalog-content\") pod \"47680b40-812e-4c78-b77f-52ff4769af25\" (UID: \"47680b40-812e-4c78-b77f-52ff4769af25\") " Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.281386 4904 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47680b40-812e-4c78-b77f-52ff4769af25-utilities\") pod \"47680b40-812e-4c78-b77f-52ff4769af25\" (UID: \"47680b40-812e-4c78-b77f-52ff4769af25\") " Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.282529 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47680b40-812e-4c78-b77f-52ff4769af25-utilities" (OuterVolumeSpecName: "utilities") pod "47680b40-812e-4c78-b77f-52ff4769af25" (UID: "47680b40-812e-4c78-b77f-52ff4769af25"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.287638 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47680b40-812e-4c78-b77f-52ff4769af25-kube-api-access-nn8bz" (OuterVolumeSpecName: "kube-api-access-nn8bz") pod "47680b40-812e-4c78-b77f-52ff4769af25" (UID: "47680b40-812e-4c78-b77f-52ff4769af25"). InnerVolumeSpecName "kube-api-access-nn8bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.323403 4904 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47680b40-812e-4c78-b77f-52ff4769af25-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "47680b40-812e-4c78-b77f-52ff4769af25" (UID: "47680b40-812e-4c78-b77f-52ff4769af25"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.382915 4904 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47680b40-812e-4c78-b77f-52ff4769af25-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.382943 4904 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47680b40-812e-4c78-b77f-52ff4769af25-utilities\") on node \"crc\" DevicePath \"\"" Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.382955 4904 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nn8bz\" (UniqueName: \"kubernetes.io/projected/47680b40-812e-4c78-b77f-52ff4769af25-kube-api-access-nn8bz\") on node \"crc\" DevicePath \"\"" Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.759494 4904 generic.go:334] "Generic (PLEG): container finished" podID="47680b40-812e-4c78-b77f-52ff4769af25" containerID="2250c184f1bb68d4c3f8294a8ea654d620e7ae336e7e0df56c7bdf9baf56356f" exitCode=0 Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.759545 4904 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q9fgk" Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.759567 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9fgk" event={"ID":"47680b40-812e-4c78-b77f-52ff4769af25","Type":"ContainerDied","Data":"2250c184f1bb68d4c3f8294a8ea654d620e7ae336e7e0df56c7bdf9baf56356f"} Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.760784 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9fgk" event={"ID":"47680b40-812e-4c78-b77f-52ff4769af25","Type":"ContainerDied","Data":"46ba01848c723dda8cf0f13c3d56c4dfdc83bcd23622101ec8dfa3b35807fff1"} Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.760809 4904 scope.go:117] "RemoveContainer" containerID="2250c184f1bb68d4c3f8294a8ea654d620e7ae336e7e0df56c7bdf9baf56356f" Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.784696 4904 scope.go:117] "RemoveContainer" containerID="8b101076b67277bf6ea094c1e09e4680362a1dc4ee0ea85ff0dacb49618d5515" Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.817560 4904 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q9fgk"] Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.823646 4904 scope.go:117] "RemoveContainer" containerID="1c9b7efc60cae2b059ced6010d5941885723257ff435db2c091617c09b336d78" Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.826653 4904 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q9fgk"] Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.849078 4904 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47680b40-812e-4c78-b77f-52ff4769af25" path="/var/lib/kubelet/pods/47680b40-812e-4c78-b77f-52ff4769af25/volumes" Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.859716 4904 scope.go:117] "RemoveContainer" containerID="2250c184f1bb68d4c3f8294a8ea654d620e7ae336e7e0df56c7bdf9baf56356f" Feb 14 12:40:57 crc kubenswrapper[4904]: E0214 12:40:57.860336 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2250c184f1bb68d4c3f8294a8ea654d620e7ae336e7e0df56c7bdf9baf56356f\": container with ID starting with 2250c184f1bb68d4c3f8294a8ea654d620e7ae336e7e0df56c7bdf9baf56356f not found: ID does not exist" containerID="2250c184f1bb68d4c3f8294a8ea654d620e7ae336e7e0df56c7bdf9baf56356f" Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.860376 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2250c184f1bb68d4c3f8294a8ea654d620e7ae336e7e0df56c7bdf9baf56356f"} err="failed to get container status \"2250c184f1bb68d4c3f8294a8ea654d620e7ae336e7e0df56c7bdf9baf56356f\": rpc error: code = NotFound desc = could not find container \"2250c184f1bb68d4c3f8294a8ea654d620e7ae336e7e0df56c7bdf9baf56356f\": container with ID starting with 2250c184f1bb68d4c3f8294a8ea654d620e7ae336e7e0df56c7bdf9baf56356f not found: ID does not exist" Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.860402 4904 scope.go:117] "RemoveContainer" containerID="8b101076b67277bf6ea094c1e09e4680362a1dc4ee0ea85ff0dacb49618d5515" Feb 14 12:40:57 crc kubenswrapper[4904]: E0214 12:40:57.860798 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b101076b67277bf6ea094c1e09e4680362a1dc4ee0ea85ff0dacb49618d5515\": container with ID starting with 8b101076b67277bf6ea094c1e09e4680362a1dc4ee0ea85ff0dacb49618d5515 not found: ID does not exist" containerID="8b101076b67277bf6ea094c1e09e4680362a1dc4ee0ea85ff0dacb49618d5515" Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.860858 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b101076b67277bf6ea094c1e09e4680362a1dc4ee0ea85ff0dacb49618d5515"} err="failed to get container status \"8b101076b67277bf6ea094c1e09e4680362a1dc4ee0ea85ff0dacb49618d5515\": rpc error: code = NotFound desc = could not find container \"8b101076b67277bf6ea094c1e09e4680362a1dc4ee0ea85ff0dacb49618d5515\": container with ID starting with 8b101076b67277bf6ea094c1e09e4680362a1dc4ee0ea85ff0dacb49618d5515 not found: ID does not exist" Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.860886 4904 scope.go:117] "RemoveContainer" containerID="1c9b7efc60cae2b059ced6010d5941885723257ff435db2c091617c09b336d78" Feb 14 12:40:57 crc kubenswrapper[4904]: E0214 12:40:57.861144 4904 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c9b7efc60cae2b059ced6010d5941885723257ff435db2c091617c09b336d78\": container with ID starting with 1c9b7efc60cae2b059ced6010d5941885723257ff435db2c091617c09b336d78 not found: ID does not exist" containerID="1c9b7efc60cae2b059ced6010d5941885723257ff435db2c091617c09b336d78" Feb 14 12:40:57 crc kubenswrapper[4904]: I0214 12:40:57.861170 4904 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c9b7efc60cae2b059ced6010d5941885723257ff435db2c091617c09b336d78"} err="failed to get container status \"1c9b7efc60cae2b059ced6010d5941885723257ff435db2c091617c09b336d78\": rpc error: code = NotFound desc = could not find container \"1c9b7efc60cae2b059ced6010d5941885723257ff435db2c091617c09b336d78\": container with ID starting with 1c9b7efc60cae2b059ced6010d5941885723257ff435db2c091617c09b336d78 not found: ID does not exist" Feb 14 12:41:16 crc kubenswrapper[4904]: I0214 12:41:16.382980 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:41:16 crc kubenswrapper[4904]: I0214 12:41:16.383451 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:41:46 crc kubenswrapper[4904]: I0214 12:41:46.383360 4904 patch_prober.go:28] interesting pod/machine-config-daemon-tnf2t container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 14 12:41:46 crc kubenswrapper[4904]: I0214 12:41:46.383868 4904 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 14 12:41:46 crc kubenswrapper[4904]: I0214 12:41:46.383920 4904 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" Feb 14 12:41:46 crc kubenswrapper[4904]: I0214 12:41:46.384614 4904 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"46ebab2bf7a2f57a967bd380510e9979599c11a1b684125f1f852ebbd4f4e829"} pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 14 12:41:46 crc kubenswrapper[4904]: I0214 12:41:46.384665 4904 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" podUID="7790d276-2506-4641-8c22-67e6441417f0" containerName="machine-config-daemon" containerID="cri-o://46ebab2bf7a2f57a967bd380510e9979599c11a1b684125f1f852ebbd4f4e829" gracePeriod=600 Feb 14 12:41:47 crc kubenswrapper[4904]: I0214 12:41:47.205687 4904 generic.go:334] "Generic (PLEG): container finished" podID="7790d276-2506-4641-8c22-67e6441417f0" containerID="46ebab2bf7a2f57a967bd380510e9979599c11a1b684125f1f852ebbd4f4e829" exitCode=0 Feb 14 12:41:47 crc kubenswrapper[4904]: I0214 12:41:47.205763 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerDied","Data":"46ebab2bf7a2f57a967bd380510e9979599c11a1b684125f1f852ebbd4f4e829"} Feb 14 12:41:47 crc kubenswrapper[4904]: I0214 12:41:47.206127 4904 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-tnf2t" event={"ID":"7790d276-2506-4641-8c22-67e6441417f0","Type":"ContainerStarted","Data":"b52ce5e54c04b69b0262b3e216404389441f47f03ec25afc130cc2a882751c8c"} Feb 14 12:41:47 crc kubenswrapper[4904]: I0214 12:41:47.206155 4904 scope.go:117] "RemoveContainer" containerID="ac036e48b848681f2b02af57685f1505e8696601eae3e7f38e11d94c83112658" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515144066636024460 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015144066636017375 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015144053422016505 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015144053423015456 5ustar corecore